You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: _posts/2025-05-02-pt-korea-user-group-recap.md
+9-9Lines changed: 9 additions & 9 deletions
Original file line number
Diff line number
Diff line change
@@ -7,7 +7,7 @@ author: 'Jiho Kim, PyTorch Korea User Group'
7
7
At the end of March, the PyTorch Korea User Group hosted a special meetup that brought together prominent speakers for deep discussions on the PyTorch core and its broader ecosystem. With the event more than doubling in size compared to past gatherings, we were able to connect with even more developers and share insights. Huge thanks to [goorm](https://goorm.co/) for sponsoring the fantastic venue! 😄
8
8
9
9
10
-
{:style="width:100%"}
10
+
{:style="width:100%"}
11
11
12
12
13
13
@@ -20,15 +20,15 @@ This recap is for those who couldn’t attend in person, as well as for particip
20
20
Representing the PyTorch Foundation, part of the Linux Foundation, Jaeung provided an overview of how PyTorch is driving core open source technologies forward. He shared PyTorch's growth story, the many global projects currently in motion, and the ecosystem’s impressive 20%+ annual growth. The session also covered how the foundation operates, how member organizations are involved, and upcoming plans that are particularly useful for practitioners.
21
21
22
22
23
-
{:style="width:100%"}
23
+
{:style="width:100%"}
24
24
25
25
26
26
## 2️⃣ Alban Desmaison | PyTorch Roadmap
27
27
28
28
Alban shared the design philosophy behind PyTorch and Meta’s official contribution roadmap ([link](https://dev-discuss.pytorch.org/t/meta-pytorch-team-2025-h1-roadmaps/2794)). He provided a deep technical dive into the differences between Eager and Compiled modes, especially breaking down the backend architecture of device Eager execution. Practical tools and improvements were also introduced—such as memory profilers, enhanced custom operator support, and pinned memory optimizations.
29
29
30
30
31
-
{:style="width:100%"}
31
+
{:style="width:100%"}
32
32
33
33
34
34
@@ -37,15 +37,15 @@ Alban shared the design philosophy behind PyTorch and Meta’s official contribu
37
37
38
38
Rebellions is building runtime integration for their proprietary NPU architecture, fully aligned with the structural changes in PyTorch 2.0. This talk introduced the performance and scalability of their upcoming chip, their integration strategy with the PyTorch runtime, and challenges in supporting Eager Mode. Hongseok also previewed their roadmap toward releasing these features within the year.
39
39
40
-
{:style="width:100%"}
40
+
{:style="width:100%"}
41
41
42
42
43
43
44
44
## 4️⃣ Kyujin Cho | Backend.AI: A Unified Platform for All AI Accelerators
45
45
46
46
Backend.AI abstracts and integrates various AI accelerators into a unified workflow. As the diversity of accelerator architectures grows, the need for portability and infrastructure unification becomes even more important. This session showcased features across development and operations—from NPU scheduling and resource allocation to monitoring. Backend.AI currently supports accelerators from NVIDIA, Intel, Tenstorrent, Rebellions, and more.
47
47
48
-
{:style="width:100%"}
48
+
{:style="width:100%"}
49
49
50
50
51
51
@@ -54,30 +54,30 @@ Backend.AI abstracts and integrates various AI accelerators into a unified workf
54
54
This talk focused on the challenges of inference in real-world industrial applications of AI models. As new state-of-the-art models emerge rapidly, there’s a growing need for environments that can quickly validate device compatibility—ideally with one-click ease. NetsPresso is actively working on a static graph representation compatible with PyTorch, offering efficient support for model development, optimization, and testing.
55
55
56
56
57
-
{:style="width:100%"}
57
+
{:style="width:100%"}
58
58
59
59
60
60
## 6️⃣ Jungyeop Lee | The Journey to Reproduce Deepseek-R1
61
61
62
62
Jungyeop took us through his journey of reproducing Deepseek, a large language model—an effort that involved 201 experiments. He shared real-world lessons from training with Korean data, tokenizer modifications, and fine-tuning strategies. His practical insights and next steps were especially valuable for those building or re-implementing large models from scratch.
63
63
64
64
65
-
{:style="width:100%"}
65
+
{:style="width:100%"}
66
66
67
67
68
68
## 7️⃣ Sol Kim | A journey from TCP architecture to production-level LLMs
69
69
70
70
Sol presented an integrated optimization approach to deploying large models using the TCP(Tensor Contraction Processor) architecture, which supports tensor contraction at the hardware level. The talk highlighted optimization techniques built on hardware abstraction layers (HALs) and bottom-up integration strategies with PyTorch—offering a hybrid hardware-software perspective.
71
71
72
72
73
-
{:style="width:100%"}
73
+
{:style="width:100%"}
74
74
75
75
## 💡 Panel Talk & Q&A 💡
76
76
77
77
The event wrapped up with an engaging panel discussion. Attendees asked sharp questions, and the speakers offered insightful answers. It was a powerful moment that captured the community’s enthusiasm for PyTorch and their hunger for deeper technical understanding.
78
78
79
79
80
-
{:style="width:100%"}
80
+
{:style="width:100%"}
0 commit comments