This media is not supported in your browser
VIEW IN TELEGRAM
Jointly announcing EAGLE-3 with SGLang: Setting a new record in LLM inference acceleration!
- 5x🚀than vanilla (on HF)
- 1.4x🚀than EAGLE-2 (on HF)
- A record of ~400 TPS on LLama 3.1 8B with a single H100 (on SGLang)
- 1.65x🚀in latency even for large bs=64 (on SGLang)
- A new scaling law: more training data, better speedup
- Apache 2.0
Paper: https://arxiv.org/abs/2503.01840
Code: https://github.com/SafeAILab/EAGLE
SGLang version: https://github.com/sgl-project/sglang/pull/4247
@Machine_learn
- 5x🚀than vanilla (on HF)
- 1.4x🚀than EAGLE-2 (on HF)
- A record of ~400 TPS on LLama 3.1 8B with a single H100 (on SGLang)
- 1.65x🚀in latency even for large bs=64 (on SGLang)
- A new scaling law: more training data, better speedup
- Apache 2.0
Paper: https://arxiv.org/abs/2503.01840
Code: https://github.com/SafeAILab/EAGLE
SGLang version: https://github.com/sgl-project/sglang/pull/4247
@Machine_learn
Executable Code Actions Elicit Better LLM Agents
1 Feb 2024 · Xingyao Wang, Yangyi Chen, Lifan Yuan, Yizhe Zhang, Yunzhu Li, Hao Peng, Heng Ji
Large Language Model (LLM) agents, capable of performing a broad range of actions, such as invoking tools and controlling robots, show great potential in tackling real-world challenges. LLM agents are typically prompted to produce actions by generating #JSON or text in a pre-defined format, which is usually limited by constrained action space (e.g., the scope of pre-defined tools) and restricted flexibility (e.g., inability to compose multiple tools). This work proposes to use executable Python code to consolidate LLM agents' actions into a unified action space (CodeAct). Integrated with a Python interpreter, CodeAct can execute code actions and dynamically revise prior actions or emit new actions upon new observations through multi-turn interactions. Our extensive analysis of 17 LLMs on API-Bank and a newly curated benchmark shows that CodeAct outperforms widely used alternatives (up to 20% higher success rate). The encouraging performance of CodeAct motivates us to build an open-source #LLM agent that interacts with environments by executing interpretable code and collaborates with users using natural language. To this end, we collect an instruction-tuning dataset CodeActInstruct that consists of 7k multi-turn interactions using CodeAct. We show that it can be used with existing data to improve models in agent-oriented tasks without compromising their general capability. CodeActAgent, finetuned from Llama2 and Mistral, is integrated with #Python interpreter and uniquely tailored to perform sophisticated tasks (e.g., model training) using existing libraries and autonomously self-debug.
Paper: https://arxiv.org/pdf/2402.01030v4.pdf
Codes:
https://github.com/epfllm/megatron-llm
https://github.com/xingyaoww/code-act
Datasets: MMLU - GSM8K - HumanEval - MATH
@Machine_learn
1 Feb 2024 · Xingyao Wang, Yangyi Chen, Lifan Yuan, Yizhe Zhang, Yunzhu Li, Hao Peng, Heng Ji
Large Language Model (LLM) agents, capable of performing a broad range of actions, such as invoking tools and controlling robots, show great potential in tackling real-world challenges. LLM agents are typically prompted to produce actions by generating #JSON or text in a pre-defined format, which is usually limited by constrained action space (e.g., the scope of pre-defined tools) and restricted flexibility (e.g., inability to compose multiple tools). This work proposes to use executable Python code to consolidate LLM agents' actions into a unified action space (CodeAct). Integrated with a Python interpreter, CodeAct can execute code actions and dynamically revise prior actions or emit new actions upon new observations through multi-turn interactions. Our extensive analysis of 17 LLMs on API-Bank and a newly curated benchmark shows that CodeAct outperforms widely used alternatives (up to 20% higher success rate). The encouraging performance of CodeAct motivates us to build an open-source #LLM agent that interacts with environments by executing interpretable code and collaborates with users using natural language. To this end, we collect an instruction-tuning dataset CodeActInstruct that consists of 7k multi-turn interactions using CodeAct. We show that it can be used with existing data to improve models in agent-oriented tasks without compromising their general capability. CodeActAgent, finetuned from Llama2 and Mistral, is integrated with #Python interpreter and uniquely tailored to perform sophisticated tasks (e.g., model training) using existing libraries and autonomously self-debug.
Paper: https://arxiv.org/pdf/2402.01030v4.pdf
Codes:
https://github.com/epfllm/megatron-llm
https://github.com/xingyaoww/code-act
Datasets: MMLU - GSM8K - HumanEval - MATH
@Machine_learn
PiEEG kit - bioscience Lab in home for your Brain and Body
🖥 Github: https://github.com/pieeg-club/PiEEG_Kit
📕 Paper: https://arxiv.org/abs/2503.13482
🌟 Methods: https://paperswithcode.com/task/eeg-1
@Machine_learn
🌟 Methods: https://paperswithcode.com/task/eeg-1
@Machine_learn
Please open Telegram to view this post
VIEW IN TELEGRAM
Introduction to Graph Neural Networks: A Starting Point for Machine Learning Engineers
📓 Paper
@Machine_learn
📓 Paper
@Machine_learn
Forwarded from Papers
با عرض سلام
در ادامه ی کار تحقیقاتی یک مقاله مروری در حوزه پاتولوژی رو می خواهیم بنویسیم. دوستانی که مایل هستن نفرات ۲ تا ۵ این موضوع رو می تونن شرکت کنن.
✅ زمان تقریبی شروع ۲۰ فروردین.
Journal: scientific reports https://www.nature.com/srep/
Price:
2: 400$
3: 300$
4: 200$
5: 150$
توضیحات کامل و نحوه نگارش هر بخش رو خودم کمک میکنم.
@Raminmousa
@Machine_learn
@Paper4money
در ادامه ی کار تحقیقاتی یک مقاله مروری در حوزه پاتولوژی رو می خواهیم بنویسیم. دوستانی که مایل هستن نفرات ۲ تا ۵ این موضوع رو می تونن شرکت کنن.
Journal: scientific reports https://www.nature.com/srep/
Price:
2: 400$
3: 300$
4: 200$
5: 150$
توضیحات کامل و نحوه نگارش هر بخش رو خودم کمک میکنم.
@Raminmousa
@Machine_learn
@Paper4money
Please open Telegram to view this post
VIEW IN TELEGRAM
Nature
Scientific Reports
Scientific Reports publishes original research in all areas of the natural and clinical sciences. We believe that if your research is scientifically valid and ...
Machine learning books and papers
با عرض سلام در ادامه ی کار تحقیقاتی یک مقاله مروری در حوزه پاتولوژی رو می خواهیم بنویسیم. دوستانی که مایل هستن نفرات ۲ تا ۵ این موضوع رو می تونن شرکت کنن. ✅ زمان تقریبی شروع ۲۰ فروردین. Journal: scientific reports https://www.nature.com/srep/ Price: …
تنها نفر ۲ و ۳ از این باقی موندن....!
Forwarded from Papers
با عرض سلام
در ادامه ی کار تحقیقاتی یک مقاله مروری در حوزه پاتولوژی رو می خواهیم بنویسیم. دوستانی که مایل هستن نفرات ۲ و ٣ این موضوع رو می تونن شرکت کنن.
✅ زمان تقریبی شروع ۲۰ فروردین.
Journal: scientific reports https://www.nature.com/srep/
Price:
2: 400$
3: 300$
توضیحات کامل و نحوه نگارش هر بخش رو خودم کمک میکنم.
@Raminmousa
@Machine_learn
@Paper4money
در ادامه ی کار تحقیقاتی یک مقاله مروری در حوزه پاتولوژی رو می خواهیم بنویسیم. دوستانی که مایل هستن نفرات ۲ و ٣ این موضوع رو می تونن شرکت کنن.
Journal: scientific reports https://www.nature.com/srep/
Price:
2: 400$
3: 300$
توضیحات کامل و نحوه نگارش هر بخش رو خودم کمک میکنم.
@Raminmousa
@Machine_learn
@Paper4money
Please open Telegram to view this post
VIEW IN TELEGRAM
Nature
Scientific Reports
Scientific Reports publishes original research in all areas of the natural and clinical sciences. We believe that if your research is scientifically valid and ...
Seg-Zero: Reasoning-Chain Guided Segmentation via Cognitive Reinforcement
🖥 Github: https://github.com/dvlab-research/Seg-Zero
📕 Paper: https://arxiv.org/abs/2503.06520v1
🌟 Dataset: https://paperswithcode.com/dataset/refcoco
📌 Model: https://huggingface.co/Ricky06662/Seg-Zero-7B
@Machine_learn
🌟 Dataset: https://paperswithcode.com/dataset/refcoco
📌 Model: https://huggingface.co/Ricky06662/Seg-Zero-7B
@Machine_learn
Please open Telegram to view this post
VIEW IN TELEGRAM
Please open Telegram to view this post
VIEW IN TELEGRAM
Greetings.
As part of our research, we want to write a review article in the field of pathology. Friends who are interested in the 2nd and 3rd places on this topic can participate.
✅ Approximate start time: April 10th.
Journal: scientific reports https://www.nature.com/srep/
Price:
2: $400
3: $300
I will help with complete explanations and how to write each section.
@Raminmousa
@Machine_learn
@Paper4money
As part of our research, we want to write a review article in the field of pathology. Friends who are interested in the 2nd and 3rd places on this topic can participate.
✅ Approximate start time: April 10th.
Journal: scientific reports https://www.nature.com/srep/
Price:
2: $400
3: $300
I will help with complete explanations and how to write each section.
@Raminmousa
@Machine_learn
@Paper4money
Nature
Scientific Reports
Scientific Reports publishes original research in all areas of the natural and clinical sciences. We believe that if your research is scientifically valid and ...
FastCuRL: Curriculum Reinforcement Learning with Progressive Context Extension for Efficient Training R1-like Reasoning Models
🖥 Github: https://github.com/nick7nlp/FastCuRL
📕 Paper: https://arxiv.org/abs/2503.17287v1
🌟 Tasks: https://paperswithcode.com/task/language-modeling
@Machine_learn
🌟 Tasks: https://paperswithcode.com/task/language-modeling
@Machine_learn
Please open Telegram to view this post
VIEW IN TELEGRAM
This media is not supported in your browser
VIEW IN TELEGRAM
Magic of open source is taking over the Video LoRA space✨
just dropped👇🔥
🍬LTX video community LoRA trainer with I2V support
🍬LTX video Cakify LoRA
🍬LTX video Squish LoRA
(🧨diffusers & comfy workflow)
trainer: https://github.com/Lightricks/LTX-Video-Trainer
LoRA: https://huggingface.co/Lightricks/LTX-Video-Cakeify-LoRA
LoRA2 : https://huggingface.co/Lightricks/LTX-Video-Squish-LoRA
🔥
@Machine_learn
just dropped👇🔥
🍬LTX video community LoRA trainer with I2V support
🍬LTX video Cakify LoRA
🍬LTX video Squish LoRA
(🧨diffusers & comfy workflow)
trainer: https://github.com/Lightricks/LTX-Video-Trainer
LoRA: https://huggingface.co/Lightricks/LTX-Video-Cakeify-LoRA
LoRA2 : https://huggingface.co/Lightricks/LTX-Video-Squish-LoRA
@Machine_learn
Please open Telegram to view this post
VIEW IN TELEGRAM