Activity
Mon
Wed
Fri
Sun
Aug
Sep
Oct
Nov
Dec
Jan
Feb
Mar
Apr
May
Jun
Jul
What is this?
Less
More

Memberships

Data Alchemy

35.8k members • Free

Advanced Data Science Society

115 members • $10/m

1184 contributions to Data Alchemy
Do You Really Need a Foundation Model?
LLM or custom model: how should you choose the right solution? https://towardsdatascience.com/do-you-really-need-a-foundation-model/
0
0
Awesome-RAG-Reasoning
A curated collection of resources, papers, tools, and implementations that bridge the gap between Retrieval-Augmented Generation (RAG) and Reasoning in Large Language Models and Agents. This repository brings together traditionally separate research domains to enable more powerful Agentic AI systems. https://github.com/DavidZWZ/Awesome-RAG-Reasoning
0
0
PageSpace is an AI-native knowledge base
PageSpace is an AI-native knowledge base that transforms your scattered documents, conversations, and project files into a single, intelligent, and perpetually searchable engine. https://github.com/2witstudios/pagespace.team
0
0
Empirical evidence of LLM’s influence on human spoken communication
"From the invention of writing and the printing press, to television and social media, human history is punctuated by major innovations in communication technology, which fundamentally altered how ideas spread and reshaped our culture. Recent chatbots powered by generative artificial intelligence constitute a novel medium that encodes cultural patterns in their neural representations and disseminates them in conversations with hundreds of millions of people. Understanding whether these patterns transmit into human language, and ultimately shape human culture, is a fundamental question. While fully quantifying the causal impact of a chatbot like ChatGPT on human culture is very challenging, lexicographic shift in human spoken communication may offer an early indicator of such broad phenomenon. Here, we apply econometric causal inference techniques [10] to 740,249 hours of human discourse from 360,445 YouTube academic talks and 771,591 conversational podcast episodes across multiple disciplines. We detect a measurable and abrupt increase in the use of words preferentially generated by ChatGPT—such as delve, comprehend, boast, swift, and meticulous—after its release. These findings suggest a scenario where machines, originally trained on human data and subsequently exhibiting their own cultural traits, can, in turn, measurably reshape human culture. This marks the beginning of a closed cultural feedback loop in which cultural traits circulate bidirectionally between humans and machines. Our results motivate further research into the evolution of human-machine culture, and raise concerns over the erosion of linguistic and cultural diversity, and the risks of scalable manipulation." https://arxiv.org/pdf/2409.01754
Gemma T5Gemma: A new collection of encoder-decoder Gemma models
"In the rapidly evolving landscape of large language models (LLMs), the spotlight has largely focused on the decoder-only architecture. While these models have shown impressive capabilities across a wide range of generation tasks, the classic encoder-decoder architecture, such as T5 (The Text-to-Text Transfer Transformer), remains a popular choice for many real-world applications. Encoder-decoder models often excel at summarization, translation, QA, and more due to their high inference efficiency, design flexibility, and richer encoder representation for understanding input. Nevertheless, the powerful encoder-decoder architecture has received little relative attention. Today, we revisit this architecture and introduce T5Gemma, a new collection of encoder-decoder LLMs developed by converting pretrained decoder-only models into the encoder-decoder architecture through a technique called adaptation. T5Gemma is based on the Gemma 2 framework, including adapted Gemma 2 2B and 9B models as well as a set of newly trained T5-sized models (Small, Base, Large and XL). We are excited to release pretrained and instruction-tuned T5Gemma models to the community to unlock new opportunities for research and development." https://developers.googleblog.com/en/t5gemma/
1
0
1-10 of 1,184
Marcio Pacheco
7
1,267points to level up
@marcio-pacheco-6005
Tech & Advertising Entrepreneur based in Seattle.

Active 2h ago
Joined Jan 24, 2024
Seattle, WA USA
Powered by