Sean (Xiang) Ren
Los Angeles, California, United States
3K followers
500 connections
About
Expose large models to complex reasoning challenges; unlock model’s potentials to…
Services
Experience
Education
View Sean (Xiang)’s full profile
Other similar profiles
-
Jiajun Wu
Stanford, CAConnect -
Pablo Rivas
Waco, TXConnect -
Carl Vondrick
New York, NYConnect -
Mohamed Elhoseiny
San Francisco Bay AreaConnect -
Srijan Kumar, Ph.D.
San Francisco, CAConnect -
Mo Sarwat
San Francisco, CAConnect -
Karthik Narasimhan
United StatesConnect -
Alfredo Canziani
New York City Metropolitan AreaConnect -
Jelani Nelson
Berkeley, CAConnect -
Negin Forouzesh
Assistant Professor at California State University, Los Angeles
United StatesConnect -
Min Xu
Pittsburgh, PAConnect -
Mohit Bansal
Parker Distinguished Professor, Computer Science Department, UNC Chapel Hill at University of North Carolina at Chapel Hill
Raleigh-Durham-Chapel Hill AreaConnect -
Jure Leskovec
Stanford, CAConnect -
Aniket Bera
West Lafayette, INConnect -
Subhabrata (Subho) Mukherjee
Greater Seattle AreaConnect -
Amir Zamir
United StatesConnect -
Sindhu Ernala
San Francisco Bay AreaConnect -
Mansi Khemka
United StatesConnect -
Abhronil Sengupta
Joseph R. and Janice M. Monkowski Career Development Associate Professor of EECS at Penn State University
State College, PAConnect -
Franz Och
Mountain View, CAConnect
Explore more posts
-
Sudarshan Lamkhede
The BayLearn 2024 abstract submission deadline has been extended to Aug 5, 2024. If you are working on cutting-edge ML research, please consider submitting an abstract. BayLearn 2024 will be an in-person event, held on Oct 10, 2024: https://baylearn.org This is a great opportunity for student researchers to showcase their latest research work and get good feedback. Submissions are non-archival and the review process is quite informal. #machinelearning #research
69 -
Manish Sainani 🤫
Great work here Rohan Sidankar . Thanks to Rohan and Justin, hushh is known for its work in optimizing neural network search through dimensionality reduction, specifically with ViT-B-32 CLIP for various search tasks as we build out our search index. This allows for maintaining recall while reducing vector size, enhancing search efficiency and effectiveness against large datasets like the LAION Benchmark Repository. Dimensionality reduction impacts neural network search by enabling the maintenance of recall while reducing vector size. This optimization, as seen in Rohan Sidankar's work at Hushh with ViT-B-32 CLIP, enhances search efficiency and effectiveness against large datasets like the LAION Benchmark Repository. ViT-B-32 CLIP can enhance search tasks by allowing for dimensionality reduction in neural network search, as demonstrated by Rohan Sidankar's work at Hushh. This approach enables the maintenance of recall while reducing the vector size, which is crucial for efficient and effective search tasks against large datasets like the LAION Benchmark Repository.
11 -
Nick Tarazona, MD
👉🏼 Evaluating the strengths and weaknesses of large language models in answering neurophysiology questions 🤓 Hassan Shojaee-Mend 👇🏻 https://lnkd.in/eZEbA_zF 🔍 Focus on data insights: - Large language models (LLMs) showcase remarkable natural language processing capabilities. - Evaluation of LLMs' proficiency in neurophysiology is crucial for research, education, and clinical applications. - Assessment of LLMs in answering neurophysiology questions in English and Persian (Farsi) reveals insights into their utility. 💡 Main outcomes and implications: - LLMs demonstrate good performance in answering neurophysiology questions. - No significant difference observed between English and Persian languages or cognitive levels. - Models excel in motor system topics but show weaknesses in integrative topics. 📚 Field significance: - Understanding LLMs' capabilities and limitations in neurophysiology research. - Highlighting the need for targeted training to address knowledge gaps and enhance reasoning skills. - Importance of domain-specific assessments for evaluating advancements in LLMs' performance. 🗄️: [#naturallanguageprocessing #neurophysiology #LLMs #research #datainsights]
-
Alexandra Neagu
I recently completed a detailed review of the paper "ALBERT: A Lite BERT for Self-supervised Learning of Language Representations" by Lan et al.! This fascinating study presents an innovative approach to language modelling, emphasizing efficiency and effectiveness through a lighter, more parameter-efficient architecture compared to traditional models like BERT. In my review, I delved into the strengths and weaknesses of the paper, discussing its impact and potential areas for improvement. This work has made a significant contribution to the field of natural language processing and has already influenced many advancements in the area. You can read my full review of the paper here: https://lnkd.in/ehxm5jXt #NLP #MachineLearning #ArtificialIntelligence #AI #Research #TechReview #DeepLearning
161 Comment -
Alfredo Canziani
I wrote two blog posts about SN, Leon Bottou and Yann LeCun's 1988 𝘚𝘪𝘮𝘶𝘭𝘢𝘵𝘦𝘶𝘳 𝘥𝘦 𝘕𝘦𝘶𝘳𝘰𝘯𝘦𝘴. One is an English translation of the original paper, for which I've reproduced the figures. The other is a tutorial on how to run their code on Apple silicon. https://lnkd.in/dMFxMZeT
762 Comments -
Thomas Gilbertson
This approach could turn the Generative AI industry upside down. Most everyone knows that AI requires computers to do a lot of calculations and especially for Large Language Models (LLMs) like ChatGPT. The majority of these calculations are actually “Matrix Multiplications”. (You remember, a 3x3 matrix times another 3x3 matrix). If the researchers in the article are proven correct, they have demonstrated a way that Matrix Multiplication can be replaced with a more efficient alternative inside the LLM getting nearly the same results. The article positions this savings as a win for the environment (yeah environment), but I think they are missing the boat. More efficient means less electricity and potentially a smaller need for GPU resources, the bread and butter of stock darling Nvidia. When you combine this cost saving with reduced privacy concerns of hosted LLMs (Open AI) and innovative LLM licensing schemes (LibreChat - https://lnkd.in/gKQjMfz5) you can see the cracks in the current direction of the LLM industry. Q: If you could run your own LLM on your own private environment with MIT open-source licensing and without expensive GPUs, why exactly do you need Nvidia, Open AI and alike? Bold prediction: if the researchers get this working at scale, we could see a rapid shift and commoditization in the LLM and generative AI space. All the pieces are in place. TLDR details from the article, describing the innovation: “The researchers' approach involves two main innovations: first, they created a custom LLM and constrained it to use only ternary values (-1, 0, 1) instead of traditional floating-point numbers, which allows for simpler computations. Second, the researchers redesigned the computationally expensive self-attention mechanism in traditional language models with a simpler, more efficient unit (that they called a MatMul-free Linear Gated Recurrent Unit—or MLGRU) that processes words sequentially using basic arithmetic operations instead of matrix multiplications.” #AI #GenerativeAI #Innovation
24 -
Santhosh M P.
🎉 Excited to Share a New Achievement! 🎉 I’m thrilled to have earned the NVIDIA Generative AI LLMs (NCA-GENL) Certification! 🚀 This certification validates my foundational understanding of developing, integrating, and maintaining AI-driven applications using Generative AI and Large Language Models (LLMs) with NVIDIA solutions. Over the past few months, I’ve delved deep into a variety of resources, including courses and webinars on topics such as Hugging Face's NLP, Building RAG Agents for LLMs, Diffusion Models, DeepLearning.AI's Prompt Engineering, and LangChain for LLM Application Development, to name a few. The journey has been challenging, but incredibly rewarding. I’m particularly excited about the knowledge I’ve gained regarding NVIDIA’s AI tech stack, including DGX, AI Enterprise, NeMo, NIM, and TensorRT. This certification not only strengthens my understanding of the current AI landscape but also equips me with the skills to contribute to developing cutting-edge AI applications. A big thank you to all the amazing resources and communities that helped me along the way. I look forward to applying this knowledge to future projects and continuing to grow in this exciting field! #NVIDIA #GenerativeAI #LLMs #Certification #AI #DeepLearning #MachineLearning #LangChain #HuggingFace
563 Comments
Explore collaborative articles
We’re unlocking community knowledge in a new way. Experts add insights directly into each article, started with the help of AI.
Explore MoreOthers named Sean (Xiang) Ren
1 other named Sean (Xiang) Ren is on LinkedIn
See others named Sean (Xiang) Ren