InfoQ Homepage Artificial Intelligence Content on InfoQ
-
Hugging Face Releases StarCoder, the Next-Generation LLM for Seamless Code Generation
Hugging Face and ServiceNow have partnered to develop StarCoder, a new open-source language model for code. The model created as a part of the BigCode initiative is an improved version of the StarCoderBase model trained on 35 billion Python tokens. StarCoder is a free AI code-generating system alternative to GitHub’s Copilot, DeepMind’s AlphaCode, and Amazon’s CodeWhisperer.
-
Meta Open-Sourced AI Tool to Animate Child and Amateur Drawings of Human Figure
Based on a joint research by Meta AI Research, Tencent America, MIT CSAIL, and Carnegie Mellon, Meta released Animated Drawings, an AI-based tool to create animations from hand drawn human-like characters.
-
AI, ML & Data News Roundup: HuggingGPT, AWS Bedrock, Microsoft Visual ChatGPT, and StableLM
The latest update for the week of April 17th, 2023, includes the recent advancements and announcements in the domains of data science, machine learning, and artificial intelligence. The spotlight this week is on top players like Hugging Face, AWS, Microsoft, and Stability AI, who have introduced groundbreaking innovations.
-
AWS Enters the Generative AI Race with Bedrock and Titan Foundation Models
AWS announced their entry into the generative AI race with the launch of Amazon Bedrock and Titan foundation models. Amazon aims to democratize access to generative AI technology, catering to customers across various industries and use cases. This groundbreaking development positions Amazon as a formidable competitor in the rapidly growing AI market.
-
Meta AI Introduces the Segment Anything Model, a Game-Changing Model for Object Segmentation
Meta AI has introduced the Segment Anything Model (SAM), aiming to democratize image segmentation by introducing a new task, dataset, and model. The project features the Segment Anything Model (SAM) and the Segment Anything 1-Billion mask dataset (SA-1B), which is the most extensive segmentation dataset to date.
-
Bloomberg Unveils a GPT Finance-Focused AI Model
Bloomberg has released BloombergGPT, a new large language model (LLM) that has been trained on enormous amounts of financial data and can help with a range of natural language processing (NLP) activities for the financial sector.
-
Strategies and Principles to Scale and Evolve MLOps - at QCon London
At the QCon London conference, Hien Luu, senior engineering manager for the Machine Learning Platform at DoorDash, discussed strategies and principles for scaling and evolving MLOps. With 85% of ML projects failing, understanding MLOps at an engineering level is crucial. Luu shared three core principles: "Dream Big, Start Small," "1% Better Every Day," and "Customer Obsession."
-
Responsible AI: from Principle to Practice at QCon London
At the QCon London conference, Microsoft's Mehrnoosh Sameki discussed Responsible AI principles and tools. She emphasized fairness, reliability, safety, privacy, inclusiveness, transparency, and accountability. Tools such as Fairlearn, InterpretML, and the Responsible AI dashboard help implement these principles.
-
Microsoft Semantic Kernel Enables LLM Integration with Conventional Programs
Microsoft has open sourced Semantic Kernel (SK), a lightweight SDK enabling the integration of large language models (LLMs) with conventional programs which can leverage prompt templating, vectorized memory, intelligent planning, and other capabilities.
-
The AI Revolution Is Just Getting Started: Leslie Miley Bids Us to Act Now against Its Bias and CO2
At his inaugural keynote of the QCON London conference, Leslie Miley, technical advisor for the CTO at Microsoft, spoke about AI Bias and Sustainability, and how the march towards transformative technologies, like large-scale AI and even crypto, has an inherent cost in the increased CO2 that comes with deployment at scale. More than just context and impact, he suggests mitigation techniques.
-
AWS and NVIDIA to Collaborate on Next-Gen EC2 P5 Instances for Accelerating Generative AI
AWS and NVIDIA announced the development of a highly scalable, on-demand AI infrastructure that is specifically designed for training large language models and creating advanced generative AI applications. The collaboration aims to create the most optimized and efficient system of its kind, capable of meeting the demands of increasingly complex AI tasks.
-
Stack Overflow Sentiment Survey Aims to Identify Experimental and Proven Technologies
Stack Overflow has published its 2023 Emergent Tech Sentiment Survey, aiming to gauge how developers perceive utility and impact of a range of "buzz-worthy" technologies like quantum computing, nanotechnology, blockchain, and others, as well as other more established ones including cloud computing, machine learning, robotics, and more.
-
Meta AI’s Large Language Model with 10x Fewer Parameters
Meta AI recently released a new large language model called Language Large Models Meta AI (LLaMA) that outperforms foundational models such as GPT-3 and is competitive with PaLM, despite having 10 times fewer parameters. LLaMA has better performance in language tasks such as natural questions, common-sense reasoning and mathematical reasoning.
-
Google AI Updates Universal Speech Model to Scale Automatic Speech Recognition beyond 100 Languages
Google AI has recently unveiled a new update for their Universal Speech Model (USM), to support the 1,000 Languages Initiative. The new model performs better than OpenAI Whisper for all segments of automation speech recognition.
-
Zero-Copy In-Memory Sharing of Large Distributed Data: V6d
Zero-copy and in-memory data manager Vineyard (v6d) is maintained as a CNCF sandbox project and provides distributed operators that can be utilized to share immutable data within or across cluster nodes. V6d is of interest particularly for deep network training on big (sharded) datasets such as large language and graph models.