Artwork

תוכן מסופק על ידי PyTorch, Edward Yang, and Team PyTorch. כל תוכן הפודקאסטים כולל פרקים, גרפיקה ותיאורי פודקאסטים מועלים ומסופקים ישירות על ידי PyTorch, Edward Yang, and Team PyTorch או שותף פלטפורמת הפודקאסט שלהם. אם אתה מאמין שמישהו משתמש ביצירה שלך המוגנת בזכויות יוצרים ללא רשותך, אתה יכול לעקוב אחר התהליך המתואר כאן https://he.player.fm/legal.
Player FM - אפליקציית פודקאסט
התחל במצב לא מקוון עם האפליקציה Player FM !

Mobile selective build

16:02
 
שתפו
 

Manage episode 294427807 series 2921809
תוכן מסופק על ידי PyTorch, Edward Yang, and Team PyTorch. כל תוכן הפודקאסטים כולל פרקים, גרפיקה ותיאורי פודקאסטים מועלים ומסופקים ישירות על ידי PyTorch, Edward Yang, and Team PyTorch או שותף פלטפורמת הפודקאסט שלהם. אם אתה מאמין שמישהו משתמש ביצירה שלך המוגנת בזכויות יוצרים ללא רשותך, אתה יכול לעקוב אחר התהליך המתואר כאן https://he.player.fm/legal.

What is mobile selective build? Why are we so obsessed with reducing binary size? How does selective build work? Why doesn't static linking just work? Why can't you just read out the ops used in a TorchScript model to determine what operators you actually need? What are the tradeoffs of statically determining the operator dependency graph versus tracing? What's up with the SELECTIVE_NAME macro? How the heck does selective build work at all when you have multiple mobile apps in a single Buck build system? What takeaways should I have as a regular PyTorch developer?

Further reading:

Liner notes:

  • binary size is premium; ship only what you actually need

  • big idea:

    • get the ops your model needs -> apply this to build of pytorch
  • get the ops your model needs

    • TorchScript ~> read it out directly from the model itself
    • but what if ops use other ops?
      • need a dependency graph. done with static analysis llvm (jiakai) ~> with a (possibly inaccurate) yaml checked in for easy kickstart if you don't want to run the pass (updated by bot, not operational since Feb, recommend rebuilding from scratch if you run into trouble)
    • other possibility: dynamic tracing
      • pro: no need for dependency graph, just look at what was called; works for dtypes
      • con: need representative inputs, if control flow might not cover everything
  • apply this to build of pytorch

    • ordinarily: static linking ensures stuff that isn't used gets pruned
      • but this doesn't work with distributed operator registration based on static initializers
    • how?
      • codegen - just don't generate it
      • no codegen - SELECTIVE_NAME - C++ doesn't support string in macro
    • build system integration
      • buck constraint: only one library
        • therefore: generate multiple copies of glue library
      • alt: atomize library into each operator. caffe2 used to do this; each library takes a long time to build (1m) and crashes xcode because there's too many
  • common hiccups

    • modify implementation details, some op is/isn't called anymore ~> error! usually just means some yaml needs regenerating. PyTorch Edge developers are very friendly and can help
  continue reading

83 פרקים

Artwork

Mobile selective build

PyTorch Developer Podcast

32 subscribers

published

iconשתפו
 
Manage episode 294427807 series 2921809
תוכן מסופק על ידי PyTorch, Edward Yang, and Team PyTorch. כל תוכן הפודקאסטים כולל פרקים, גרפיקה ותיאורי פודקאסטים מועלים ומסופקים ישירות על ידי PyTorch, Edward Yang, and Team PyTorch או שותף פלטפורמת הפודקאסט שלהם. אם אתה מאמין שמישהו משתמש ביצירה שלך המוגנת בזכויות יוצרים ללא רשותך, אתה יכול לעקוב אחר התהליך המתואר כאן https://he.player.fm/legal.

What is mobile selective build? Why are we so obsessed with reducing binary size? How does selective build work? Why doesn't static linking just work? Why can't you just read out the ops used in a TorchScript model to determine what operators you actually need? What are the tradeoffs of statically determining the operator dependency graph versus tracing? What's up with the SELECTIVE_NAME macro? How the heck does selective build work at all when you have multiple mobile apps in a single Buck build system? What takeaways should I have as a regular PyTorch developer?

Further reading:

Liner notes:

  • binary size is premium; ship only what you actually need

  • big idea:

    • get the ops your model needs -> apply this to build of pytorch
  • get the ops your model needs

    • TorchScript ~> read it out directly from the model itself
    • but what if ops use other ops?
      • need a dependency graph. done with static analysis llvm (jiakai) ~> with a (possibly inaccurate) yaml checked in for easy kickstart if you don't want to run the pass (updated by bot, not operational since Feb, recommend rebuilding from scratch if you run into trouble)
    • other possibility: dynamic tracing
      • pro: no need for dependency graph, just look at what was called; works for dtypes
      • con: need representative inputs, if control flow might not cover everything
  • apply this to build of pytorch

    • ordinarily: static linking ensures stuff that isn't used gets pruned
      • but this doesn't work with distributed operator registration based on static initializers
    • how?
      • codegen - just don't generate it
      • no codegen - SELECTIVE_NAME - C++ doesn't support string in macro
    • build system integration
      • buck constraint: only one library
        • therefore: generate multiple copies of glue library
      • alt: atomize library into each operator. caffe2 used to do this; each library takes a long time to build (1m) and crashes xcode because there's too many
  • common hiccups

    • modify implementation details, some op is/isn't called anymore ~> error! usually just means some yaml needs regenerating. PyTorch Edge developers are very friendly and can help
  continue reading

83 פרקים

כל הפרקים

×
 
Loading …

ברוכים הבאים אל Player FM!

Player FM סורק את האינטרנט עבור פודקאסטים באיכות גבוהה בשבילכם כדי שתהנו מהם כרגע. זה יישום הפודקאסט הטוב ביותר והוא עובד על אנדרואיד, iPhone ואינטרנט. הירשמו לסנכרון מנויים במכשירים שונים.

 

מדריך עזר מהיר

האזן לתוכנית הזו בזמן שאתה חוקר
הפעלה