π€π§ NanoChat: The Best ChatGPT That $100 Can Buy
ποΈ 20 Oct 2025
π AI News & Trends
In a world dominated by billion-dollar AI models like GPT-4 and Claude 3, itβs refreshing to see a minimalist, open-source alternative that puts the power of Large Language Models (LLMs) back into the hands of hackers, researchers and enthusiasts. Enter NanoChat β an end-to-end, full-stack implementation of a ChatGPT-style AI chatbot developed by Andrej Karpathy, ...
#NanoChat #ChatGPT #AI #LargeLanguageModels #OpenSource #AndrejKarpathy
ποΈ 20 Oct 2025
π AI News & Trends
In a world dominated by billion-dollar AI models like GPT-4 and Claude 3, itβs refreshing to see a minimalist, open-source alternative that puts the power of Large Language Models (LLMs) back into the hands of hackers, researchers and enthusiasts. Enter NanoChat β an end-to-end, full-stack implementation of a ChatGPT-style AI chatbot developed by Andrej Karpathy, ...
#NanoChat #ChatGPT #AI #LargeLanguageModels #OpenSource #AndrejKarpathy
π€π§ NanoChat: The Best ChatGPT That $100 Can Buy
ποΈ 20 Oct 2025
π AI News & Trends
In a world dominated by billion-dollar AI models like GPT-4 and Claude 3, itβs refreshing to see a minimalist, open-source alternative that puts the power of Large Language Models (LLMs) back into the hands of hackers, researchers and enthusiasts. Enter NanoChat β an end-to-end, full-stack implementation of a ChatGPT-style AI chatbot developed by Andrej Karpathy, ...
#NanoChat #ChatGPT #AI #LargeLanguageModels #OpenSource #AndrejKarpathy
ποΈ 20 Oct 2025
π AI News & Trends
In a world dominated by billion-dollar AI models like GPT-4 and Claude 3, itβs refreshing to see a minimalist, open-source alternative that puts the power of Large Language Models (LLMs) back into the hands of hackers, researchers and enthusiasts. Enter NanoChat β an end-to-end, full-stack implementation of a ChatGPT-style AI chatbot developed by Andrej Karpathy, ...
#NanoChat #ChatGPT #AI #LargeLanguageModels #OpenSource #AndrejKarpathy
π€π§ PaddleOCR-VL: Redefining Multilingual Document Parsing with a 0.9B Vision-Language Model
ποΈ 20 Oct 2025
π AI News & Trends
In an era where information is predominantly digital, the ability to extract, interpret and organize data from documents is crucial. From invoices and research papers to multilingual contracts and handwritten notes, document parsing stands at the intersection of vision and language. Traditional Optical Character Recognition (OCR) systems have made impressive strides but they often fall ...
#PaddleOCR-VL #Multilingual #DocumentParsing #VisionLanguageModel #OCR #AI
ποΈ 20 Oct 2025
π AI News & Trends
In an era where information is predominantly digital, the ability to extract, interpret and organize data from documents is crucial. From invoices and research papers to multilingual contracts and handwritten notes, document parsing stands at the intersection of vision and language. Traditional Optical Character Recognition (OCR) systems have made impressive strides but they often fall ...
#PaddleOCR-VL #Multilingual #DocumentParsing #VisionLanguageModel #OCR #AI
π€π§ PaddleOCR-VL: Redefining Multilingual Document Parsing with a 0.9B Vision-Language Model
ποΈ 20 Oct 2025
π AI News & Trends
In an era where information is predominantly digital, the ability to extract, interpret and organize data from documents is crucial. From invoices and research papers to multilingual contracts and handwritten notes, document parsing stands at the intersection of vision and language. Traditional Optical Character Recognition (OCR) systems have made impressive strides but they often fall ...
#PaddleOCR-VL #Multilingual #DocumentParsing #VisionLanguageModel #OCR #AI
ποΈ 20 Oct 2025
π AI News & Trends
In an era where information is predominantly digital, the ability to extract, interpret and organize data from documents is crucial. From invoices and research papers to multilingual contracts and handwritten notes, document parsing stands at the intersection of vision and language. Traditional Optical Character Recognition (OCR) systems have made impressive strides but they often fall ...
#PaddleOCR-VL #Multilingual #DocumentParsing #VisionLanguageModel #OCR #AI
β€1
π€π§ Top 30 More Retro Bollywood Diwali Portrait Prompts for Women Using Gemini AI β Part 2
ποΈ 20 Oct 2025
π AI News & Trends
The Diwali celebrations continue and so does the nostalgia! After the huge buzz around our Top 20 Retro Bollywood Diwali Portrait Ideas, weβre back with Part 2 featuring prompts 21 to 50 curated to help you create even more magical, cinematic AI portraits using Google Gemini AI. If you loved the 90s-style Diwali aesthetics shimmering ...
ποΈ 20 Oct 2025
π AI News & Trends
The Diwali celebrations continue and so does the nostalgia! After the huge buzz around our Top 20 Retro Bollywood Diwali Portrait Ideas, weβre back with Part 2 featuring prompts 21 to 50 curated to help you create even more magical, cinematic AI portraits using Google Gemini AI. If you loved the 90s-style Diwali aesthetics shimmering ...
β€2
Here is a trick to optimize a neural network that gives about 4x speedup when transferring data from CPU to GPU.
Let's consider an image classification task.
We define the model, load and transform the data.
In the training loop, we transfer data to the GPU and train the network.
What's the problem:
If you look into the profiler,
- most of the resources go to the kernel (i.e., the training itself),
- but a noticeable amount of time is also spent on transferring data from CPU to GPU (cudaMemcpyAsync).
This can be easily reduced.
Initially, the dataset consists of pixels as 8-bit integers. We convert them to 32-bit floats.
Then we send these float tensors to the GPU. As a result, the data size becomes 4 times larger, making the transfer heavier.
The solution:
Shift the transformation step after the transfer. That is, first transfer the 8-bit ints, and then convert them to floats on the GPU.
As a result, the data transfer step speeds up significantly.
Of course, this doesn't work everywhere; for example, in NLP we initially deal with float embeddings.
But in cases where it applies, the speedup is very noticeable.
π @DataScienceM
Let's consider an image classification task.
We define the model, load and transform the data.
In the training loop, we transfer data to the GPU and train the network.
What's the problem:
If you look into the profiler,
- most of the resources go to the kernel (i.e., the training itself),
- but a noticeable amount of time is also spent on transferring data from CPU to GPU (cudaMemcpyAsync).
This can be easily reduced.
Initially, the dataset consists of pixels as 8-bit integers. We convert them to 32-bit floats.
Then we send these float tensors to the GPU. As a result, the data size becomes 4 times larger, making the transfer heavier.
The solution:
Shift the transformation step after the transfer. That is, first transfer the 8-bit ints, and then convert them to floats on the GPU.
As a result, the data transfer step speeds up significantly.
Of course, this doesn't work everywhere; for example, in NLP we initially deal with float embeddings.
But in cases where it applies, the speedup is very noticeable.
Please open Telegram to view this post
VIEW IN TELEGRAM
β€4