#decodingaianddatascience arama sonuçları

"#decodingaianddatascience" için sonuç bulunamadı

🚨 DeepSeek just did something wild. They built an OCR system that compresses long text into vision tokens literally turning paragraphs into pixels. Their model, DeepSeek-OCR, achieves 97% decoding precision at 10× compression and still manages 60% accuracy even at 20×. That…

godofprompt's tweet image. 🚨 DeepSeek just did something wild.

They built an OCR system that compresses long text into vision tokens  literally turning paragraphs into pixels.

Their model, DeepSeek-OCR, achieves 97% decoding precision at 10× compression and still manages 60% accuracy even at 20×. That…

🚀 DeepSeek-OCR — the new frontier of OCR from @deepseek_ai , exploring optical context compression for LLMs, is running blazingly fast on vLLM ⚡ (~2500 tokens/s on A100-40G) — powered by vllm==0.8.5 for day-0 model support. 🧠 Compresses visual contexts up to 20× while keeping…

vllm_project's tweet image. 🚀 DeepSeek-OCR — the new frontier of OCR from @deepseek_ai , exploring optical context compression for LLMs, is running blazingly fast on vLLM ⚡ (~2500 tokens/s on A100-40G) — powered by vllm==0.8.5 for day-0 model support.

🧠 Compresses visual contexts up to 20× while keeping…
vllm_project's tweet image. 🚀 DeepSeek-OCR — the new frontier of OCR from @deepseek_ai , exploring optical context compression for LLMs, is running blazingly fast on vLLM ⚡ (~2500 tokens/s on A100-40G) — powered by vllm==0.8.5 for day-0 model support.

🧠 Compresses visual contexts up to 20× while keeping…
vllm_project's tweet image. 🚀 DeepSeek-OCR — the new frontier of OCR from @deepseek_ai , exploring optical context compression for LLMs, is running blazingly fast on vLLM ⚡ (~2500 tokens/s on A100-40G) — powered by vllm==0.8.5 for day-0 model support.

🧠 Compresses visual contexts up to 20× while keeping…

IT FREAKING WORKED! At 4am today I just proved DeepSeek-OCR AI can scan an entire microfiche sheet and not just cells and retain 100% of the data in seconds… AND Have a full understanding of the text/complex drawings and their context. I just changed offline data curation!

BrianRoemmele's tweet image. IT FREAKING WORKED!

At 4am today I just proved DeepSeek-OCR AI can scan an entire microfiche sheet and not just cells and retain 100% of the data in seconds…

AND

Have a full understanding of the text/complex drawings and their context.

I just changed offline data curation!

BOOOOOOOM! CHINA DEEPSEEK DOES IT AGAIN! An entire encyclopedia compressed into a single, high-resolution image! — A mind-blowing breakthrough. DeepSeek-OCR, unleashed an electrifying 3-billion-parameter vision-language model that obliterates the boundaries between text and…

BrianRoemmele's tweet image. BOOOOOOOM!

CHINA DEEPSEEK DOES IT AGAIN!

An entire encyclopedia compressed into a single, high-resolution image!

—

A mind-blowing breakthrough. DeepSeek-OCR, unleashed an electrifying 3-billion-parameter vision-language model that obliterates the boundaries between text and…


DeepSeek finally released a new model and paper. And because this DeepSeek-OCR release is a bit different from what everyone expected, and DeepSeek releases are generally a big deal, I wanted to do a brief explainer of what it is all about. In short, they explore how vision…

rasbt's tweet image. DeepSeek finally released a new model and paper. And because this DeepSeek-OCR release is a bit different from what everyone expected, and DeepSeek releases are generally a big deal, I wanted to do a brief explainer of what it is all about.

In short, they explore how vision…

BOOOOOOOM! CHINA DEEPSEEK DOES IT AGAIN! An entire encyclopedia compressed into a single, high-resolution image! — A mind-blowing breakthrough. DeepSeek-OCR, unleashed an electrifying 3-billion-parameter vision-language model that obliterates the boundaries between text and…

BrianRoemmele's tweet image. BOOOOOOOM!

CHINA DEEPSEEK DOES IT AGAIN!

An entire encyclopedia compressed into a single, high-resolution image!

—

A mind-blowing breakthrough. DeepSeek-OCR, unleashed an electrifying 3-billion-parameter vision-language model that obliterates the boundaries between text and…

A more serious thread on the DeepSeek-OCR hype / serious misinterpretation going on. 1. On token reduction via representing text in images, researchers from Cambridge have previously shown that 500x prompt token compression is possible (ACL'25, Li, Su, and Collier). Without…

Kangwook_Lee's tweet image. A more serious thread on the DeepSeek-OCR hype / serious misinterpretation going on.

1. 
On token reduction via representing text in images, researchers from Cambridge have previously shown that 500x prompt token compression is possible (ACL'25, Li, Su, and Collier). 

Without…
Kangwook_Lee's tweet image. A more serious thread on the DeepSeek-OCR hype / serious misinterpretation going on.

1. 
On token reduction via representing text in images, researchers from Cambridge have previously shown that 500x prompt token compression is possible (ACL'25, Li, Su, and Collier). 

Without…
Kangwook_Lee's tweet image. A more serious thread on the DeepSeek-OCR hype / serious misinterpretation going on.

1. 
On token reduction via representing text in images, researchers from Cambridge have previously shown that 500x prompt token compression is possible (ACL'25, Li, Su, and Collier). 

Without…

DeepSeek-OCR looks impressive, but its core idea is not new. Input “Text” as “Image” — already explored by: LANGUAGE MODELING WITH PIXELS (Phillip et al., ICLR 2023) CLIPPO: Image-and-Language Understanding from Pixels Only (Michael et al. CVPR 2023) Pix2Struct: Screenshot…

awinyimgprocess's tweet image. DeepSeek-OCR looks impressive, but its core idea is not new.

Input “Text” as “Image” — already explored by:
LANGUAGE MODELING WITH PIXELS (Phillip et al., ICLR 2023)
CLIPPO: Image-and-Language Understanding from Pixels Only (Michael et al. CVPR 2023)
Pix2Struct: Screenshot…
awinyimgprocess's tweet image. DeepSeek-OCR looks impressive, but its core idea is not new.

Input “Text” as “Image” — already explored by:
LANGUAGE MODELING WITH PIXELS (Phillip et al., ICLR 2023)
CLIPPO: Image-and-Language Understanding from Pixels Only (Michael et al. CVPR 2023)
Pix2Struct: Screenshot…
awinyimgprocess's tweet image. DeepSeek-OCR looks impressive, but its core idea is not new.

Input “Text” as “Image” — already explored by:
LANGUAGE MODELING WITH PIXELS (Phillip et al., ICLR 2023)
CLIPPO: Image-and-Language Understanding from Pixels Only (Michael et al. CVPR 2023)
Pix2Struct: Screenshot…
awinyimgprocess's tweet image. DeepSeek-OCR looks impressive, but its core idea is not new.

Input “Text” as “Image” — already explored by:
LANGUAGE MODELING WITH PIXELS (Phillip et al., ICLR 2023)
CLIPPO: Image-and-Language Understanding from Pixels Only (Michael et al. CVPR 2023)
Pix2Struct: Screenshot…

Machine Learning Image Credit : DataInterview

PythonPr's tweet image. Machine Learning
Image Credit : DataInterview

Unlike closed AI labs, DeepSeek proves they are truly open research Their OCR paper treats paragraphs as pixels and is 60x leap more efficient than traditional LLMs Small super efficient models are the future

bindureddy's tweet image. Unlike closed AI labs, DeepSeek proves they are truly open research 

Their OCR paper treats paragraphs as pixels and is  60x leap more efficient than traditional LLMs 

Small super efficient models are the future

This is the JPEG moment for AI. Optical compression doesn't just make context cheaper. It makes AI memory architectures viable. Training data bottlenecks? Solved. - 200k pages/day on ONE GPU - 33M pages/day on 20 nodes - Every multimodal model is data-constrained. Not anymore.…

RayFernando1337's tweet image. This is the JPEG moment for AI.

Optical compression doesn't just make context cheaper. It makes AI memory architectures viable.

Training data bottlenecks? Solved.
- 200k pages/day on ONE GPU
- 33M pages/day on 20 nodes
- Every multimodal model is data-constrained. Not anymore.…

OSI Model data encapsulation and de-encapsulation

thatstraw's tweet image. OSI Model data encapsulation and de-encapsulation

DeepSeek-OCR Contexts Optical Compression

_akhaliq's tweet image. DeepSeek-OCR

Contexts Optical Compression

High-resolution satellite images can be insanely expensive to buy. So here's a list of free datasets you can access. These datasets can be used to build foundation models, super-resolution models, or for segmentation.

yohaniddawela's tweet image. High-resolution satellite images can be insanely expensive to buy.

So here's a list of free datasets you can access.

These datasets can be used to build foundation models, super-resolution models, or for segmentation.

DeepSeek-OCR is out! 🔥 my take ⤵️ > pretty insane it can parse and re-render charts in HTML > it uses CLIP and SAM features concatenated, so better grounding > very efficient per vision tokens/performance ratio > covers 100 languages

mervenoyann's tweet image. DeepSeek-OCR is out! 🔥 my take ⤵️ 
> pretty insane it can parse and re-render charts in HTML
> it uses CLIP and SAM features concatenated, so better grounding
> very efficient per vision tokens/performance ratio
> covers 100 languages

DeepSeek just dropped a new OCR model! And this isn't about OCR. We've all heard "a picture is worth a thousand words." DeepSeek literally proved it. They've built a breakthrough in AI memory compression that could change how models handle long contexts. The core idea:…

akshay_pachaar's tweet image. DeepSeek just dropped a new OCR model!

And this isn't about OCR.

We've all heard "a picture is worth a thousand words." DeepSeek literally proved it.

They've built a breakthrough in AI memory compression that could change how models handle long contexts.

The core idea:…

cHiNa cAnNoT iNnOvAtE DeepSeek just revolutionized OCR - how computers read scanned documents. The traditional way is to read each word; and LLM’s create one token for each word. What DeepSeek does is to treat each word as an image. This introduces the concept of compression.…

Kanthan2030's tweet image. cHiNa cAnNoT iNnOvAtE

DeepSeek just revolutionized OCR - how computers read scanned documents.

The traditional way is to read each word; and LLM’s create one token for each word.

What DeepSeek does is to treat each word as an image. This introduces the concept of compression.…

DeepSeek released an OCR model today. Their motivation is really interesting: they want to use visual modality as an efficient compression medium for textual information, and use this to solve long-context challenges in LLMs. Of course, they are using it to get more training…

iScienceLuvr's tweet image. DeepSeek released an OCR model today. 

Their motivation is really interesting: they want to use visual modality as an efficient compression medium for textual information, and use this to solve long-context challenges in LLMs.

Of course, they are using it to get more training…

DeepSeek-OCR just dropped. 🔥 Sets a new standard for open-source OCR A 3B-parameter vision-language model designed for high-performance optical character recognition and structured document conversion. - Can parse and re-render charts in HTML - Optical Context Compression:…

rohanpaul_ai's tweet image. DeepSeek-OCR just dropped. 🔥

Sets a new standard for open-source OCR

A 3B-parameter vision-language model designed for high-performance optical character recognition and structured document conversion. 

-  Can parse and re-render charts in HTML

- Optical Context Compression:…

NEW DeepSeek OCR model that outperforms dots ocr while prefilling 3x less tokens

casper_hansen_'s tweet image. NEW DeepSeek OCR model that outperforms dots ocr while prefilling 3x less tokens

Loading...

Something went wrong.


Something went wrong.


United States Trends