techvignesh's profile picture. VP - Leading the Group AI Technology & Data Marketplace @ ABSA Bank.. Prev @ Tata Innovation Lab / Walmart / Mashreq Bank / Barclays.

Vignesh Varadharajan

@techvignesh

VP - Leading the Group AI Technology & Data Marketplace @ ABSA Bank.. Prev @ Tata Innovation Lab / Walmart / Mashreq Bank / Barclays.

Vignesh Varadharajan 已轉發

With the announcement of S3-native-streams (Freight clusters), here is a commentary on Confluent strategy regarding object storage, streaming and an open data architecture. jack-vanlightly.com/blog/2024/5/2/…


Vignesh Varadharajan 已轉發

[New survey on #Knowledge Distillation for #LLMs] 🚀 KD is key for finetuning & aligning LLMs, transferring knowledge from teacher to student models🧑‍🏫➡️🧑‍🎓. We explore KD Algorithms, and Skill & Vertical Distillation. Learn more: arxiv.org/abs/2402.13116 1/5

shawnxxh's tweet image. [New survey on #Knowledge Distillation for #LLMs] 🚀  

KD is key for finetuning & aligning LLMs, transferring knowledge from teacher to student models🧑‍🏫➡️🧑‍🎓. We explore KD Algorithms, and Skill & Vertical Distillation. 

Learn more: arxiv.org/abs/2402.13116
1/5

Vignesh Varadharajan 已轉發

Today we are announcing our latest project, an effort to provide a new evaluation system for open source models. Traditional benchmarking leans heavily on public datasets which can be easy to game and often lead to superficial score improvements that mask true model capabilities…

NousResearch's tweet image. Today we are announcing our latest project, an effort to provide a new evaluation system for open source models. Traditional benchmarking leans heavily on public datasets which can be easy to game and often lead to superficial score improvements that mask true model capabilities…

Vignesh Varadharajan 已轉發

It's been nice to see small jumps in output quality in our RAG applications from chunking experiments, contextual preprocessing, prompt engineering, fine-tuned embeddings, lexical search, reranking, etc. but we just added Mixtral-8x7B-Instruct to the mix and we're seeing a 🤯…

GokuMohandas's tweet image. It's been nice to see small jumps in output quality in our RAG applications from chunking experiments, contextual preprocessing, prompt engineering, fine-tuned embeddings, lexical search, reranking, etc. but we just added Mixtral-8x7B-Instruct to the mix and we're seeing a 🤯…
GokuMohandas's tweet image. It's been nice to see small jumps in output quality in our RAG applications from chunking experiments, contextual preprocessing, prompt engineering, fine-tuned embeddings, lexical search, reranking, etc. but we just added Mixtral-8x7B-Instruct to the mix and we're seeing a 🤯…
GokuMohandas's tweet image. It's been nice to see small jumps in output quality in our RAG applications from chunking experiments, contextual preprocessing, prompt engineering, fine-tuned embeddings, lexical search, reranking, etc. but we just added Mixtral-8x7B-Instruct to the mix and we're seeing a 🤯…
GokuMohandas's tweet image. It's been nice to see small jumps in output quality in our RAG applications from chunking experiments, contextual preprocessing, prompt engineering, fine-tuned embeddings, lexical search, reranking, etc. but we just added Mixtral-8x7B-Instruct to the mix and we're seeing a 🤯…

Vignesh Varadharajan 已轉發

If you are using LoRA or QLoRA, welcome to try the drop-in replacement LoftQ which minimizes the discrepancy between W and its quantized counterpart Q via a better LoRA B & A initialization. More research work is needed for LoRA with Quantization. PEFT: github.com/huggingface/pe…

WeizhuChen's tweet image. If you are using LoRA or QLoRA, welcome to try the drop-in replacement LoftQ which minimizes the discrepancy between W and its quantized counterpart Q via a better LoRA B & A initialization.  More research work is needed for LoRA with Quantization. PEFT: github.com/huggingface/pe…

Vignesh Varadharajan 已轉發

[Arena Update] We've collected over 6000 and 1500 votes for Mixtral-8x7B and Gemini Pro. Both show strong performance against GPT-3.5-Turbo. Big congrats again on the release! @MistralAI @GoogleDeepMind Full leaderboard: huggingface.co/spaces/lmsys/c…

arena's tweet image. [Arena Update]
We've collected over 6000 and 1500 votes for Mixtral-8x7B and Gemini Pro. Both show strong performance against GPT-3.5-Turbo.

Big congrats again on the release! @MistralAI @GoogleDeepMind

Full leaderboard: huggingface.co/spaces/lmsys/c…
arena's tweet image. [Arena Update]
We've collected over 6000 and 1500 votes for Mixtral-8x7B and Gemini Pro. Both show strong performance against GPT-3.5-Turbo.

Big congrats again on the release! @MistralAI @GoogleDeepMind

Full leaderboard: huggingface.co/spaces/lmsys/c…
arena's tweet image. [Arena Update]
We've collected over 6000 and 1500 votes for Mixtral-8x7B and Gemini Pro. Both show strong performance against GPT-3.5-Turbo.

Big congrats again on the release! @MistralAI @GoogleDeepMind

Full leaderboard: huggingface.co/spaces/lmsys/c…

♊️Gemini is now in the Arena. Excited to see its ranking with human evals! Meanwhile, our server just hit the highest traffic since May, with 10,000 votes in just 2 days. How incredible! Huge thanks to @karpathy and the amazing community😂Let's vote at chat.lmsys.org

arena's tweet image. ♊️Gemini is now in the Arena. Excited to see its ranking with human evals!

Meanwhile, our server just hit the highest traffic since May, with 10,000 votes in just 2 days. How incredible!

Huge thanks to @karpathy and the amazing community😂Let's vote at chat.lmsys.org


Vignesh Varadharajan 已轉發

📢 UltraFeedback Curated by @argilla_io After Notus, we wanna improve data quality for the OS AI community 🐛Fixed 1,968 data points with distilabel 🤯 Used the UltraFeedback method to fix the Ultrafeedback dataset. More in the 🧵 💾 Dataset: huggingface.co/datasets/argil… 🧵


Vignesh Varadharajan 已轉發

A Round-up of 20 Exciting LLM-related Papers by @seb_ruder Sebastian has done an incredible job in sifting through 3586 papers to bring us a curated selection of 20 standout #NLP papers from #NeurIPS2023 Here's a quick glimpse into the main trends that are defining the future…

pandeyparul's tweet image. A Round-up of 20 Exciting LLM-related Papers by @seb_ruder 

Sebastian has done an incredible job in sifting through 3586 papers to bring us a curated selection of 20 standout #NLP papers from #NeurIPS2023  
Here's a quick glimpse into the main trends that are defining the future…

Vignesh Varadharajan 已轉發

I've been maintaining a database of base models with detailed info about the licensing here. See the screenshot for the list of OS-licensed models sorted by size. I believe BTPM-3B /Mistral-7B / MPT-30B is "best model per VRAM" tradeoff. docs.google.com/spreadsheets/d…

BlancheMinerva's tweet image. I've been maintaining a database of base models with detailed info about the licensing here. See the screenshot for the list of OS-licensed models sorted by size. I believe BTPM-3B /Mistral-7B / MPT-30B is "best model per VRAM" tradeoff.

docs.google.com/spreadsheets/d…

Vignesh Varadharajan 已轉發

Okay, I've created an "awesome repository" that lists all the GPTs related to cybersecurity. Take a look – the list is continuously growing and there are already many use cases! Feel free to add yours 👇#gpt #infosec #Agents github.com/fr0gger/Awesom…

fr0gger_'s tweet image. Okay, I've created an "awesome repository" that lists all the GPTs related to cybersecurity. Take a look – the list is continuously growing and there are already many use cases! Feel free to add yours 👇#gpt #infosec #Agents

github.com/fr0gger/Awesom…

Vignesh Varadharajan 已轉發

How well do long-context LLMs (gpt-4-turbo, claude-2) recall specifics in BIG documents? (>= 250k tokens) Inspired by @GregKamradt’s work on stress-testing gpt-4 128k, we extended this by stress testing gpt-4/Claude on even bigger documents that overflow the context window,…

jerryjliu0's tweet image. How well do long-context LLMs (gpt-4-turbo, claude-2) recall specifics in BIG documents? (>= 250k tokens)

Inspired by @GregKamradt’s work on stress-testing gpt-4 128k, we extended this by stress testing gpt-4/Claude on even bigger documents that overflow the context window,…

Vignesh Varadharajan 已轉發

Embedding fine-tuning is underrated and underexplored. An easy trick you can do on top of any black-box embedding model (e.g. openai) is to fine-tune a query embedding transformation. Can be linear, a NN, or anything else. Optimize retrieval perf + no need to reindex docs! 👇

Previously we've seen how to improve retrieval by funetuning an embedding model. @llama_index also supports finetuning an adapter on top of existing models, which lets us improve retrieval without updating our existing embeddings. 🚀 Let's see how it works 👇🧵

rsrohan99's tweet image. Previously we've seen how to improve retrieval by funetuning an embedding model.

@llama_index also supports finetuning an adapter on top of existing models, which lets us improve retrieval without updating our existing embeddings. 🚀

Let's see how it works 👇🧵


Vignesh Varadharajan 已轉發

Did you know there are better and cheaper embedding models than @OpenAI? 🤔 We are excited to launch Text Embedding Inference (TEI) on @huggingface Inference endpoints. TEI is a purpose-built solution to run open-source embedding 🚀 👉 huggingface.co/blog/inference… 🧶


Vignesh Varadharajan 已轉發

Open Source AI repos that caught my 👀 this week @MetaGPT_ github.com/geekan/MetaGPT - multi agent collaboration - MetaGPT encodes Standard Operating Procedures (SOPs) into prompts. The claim is that it takes a one line requirement as input and outputs user stories / competitive…

oanaolt's tweet image. Open Source AI repos that caught my 👀 this week

@MetaGPT_ github.com/geekan/MetaGPT - multi agent collaboration - MetaGPT encodes Standard Operating Procedures (SOPs) into prompts. The claim is that it takes a one line requirement as input and outputs user stories / competitive…

Vignesh Varadharajan 已轉發

A high-quality LLMs watch/reading list here: gist.github.com/rain-1/eebd5e5… Contains videos/lectures/articles that do great job at explaining LLMs and GPT models including our "Transformer Blueprint" article :-)

Jeande_d's tweet image. A high-quality LLMs watch/reading list here: gist.github.com/rain-1/eebd5e5…

Contains videos/lectures/articles that do great job at explaining LLMs and GPT models including our "Transformer Blueprint" article :-)

Vignesh Varadharajan 已轉發

I just uploaded a 90 minute tutorial, which is designed to be the one place I point coders at when they ask "hey, tell me everything I need to know about LLMs!" It starts at the basics: the 3-step pre-training / fine-tuning / classifier ULMFiT approach used in all modern LLMs.

jeremyphoward's tweet image. I just uploaded a 90 minute tutorial, which is designed to be the one place I point coders at when they ask "hey, tell me everything I need to know about LLMs!"

It starts at the basics: the 3-step pre-training / fine-tuning / classifier ULMFiT approach used in all modern LLMs.

Vignesh Varadharajan 已轉發

GitWit Agent is Open Source! 🤝 It took three months to build and has been used by hundreds of users to generate over 7,000 commits on GitHub! 📥 It's also available as a command-line tool. Source code: github.com/jamesmurdza/gi…

jamesmurdza's tweet image. GitWit Agent is Open Source! 🤝

It took three months to build and has been used by hundreds of users to generate over 7,000 commits on GitHub! 📥

It's also available as a command-line tool.

Source code: github.com/jamesmurdza/gi…

Vignesh Varadharajan 已轉發

⚡️Run LLMs locally with CTranslate2 We've added support for running local models with the blazingly fast CTranslate2. Thanks to GH eryk-dsai from @deepsense_ai for the feature and @HamelHusain for the great post that introduced us to the library! Docs: python.langchain.com/docs/integrati…

LangChainAI's tweet image. ⚡️Run LLMs locally with CTranslate2

We've added support for running local models with the blazingly fast CTranslate2.

Thanks to GH eryk-dsai from @deepsense_ai for the feature and @HamelHusain for the great post that introduced us to the library!

Docs: python.langchain.com/docs/integrati…
LangChainAI's tweet image. ⚡️Run LLMs locally with CTranslate2

We've added support for running local models with the blazingly fast CTranslate2.

Thanks to GH eryk-dsai from @deepsense_ai for the feature and @HamelHusain for the great post that introduced us to the library!

Docs: python.langchain.com/docs/integrati…

I updated this thanks to @abacaj CTranslate2 allows you to squeeze significantly more performance beyond vLLM

HamelHusain's tweet image. I updated this thanks to @abacaj   CTranslate2 allows you to squeeze significantly more performance beyond vLLM


Loading...

Something went wrong.


Something went wrong.