Stability AI’s Stable Audio 2.0 – A New Era in AI-Generated Music

Cohere launches Command R+ on Microsoft Azure, a revolutionary enterprise LLM that beats GPT-4 Turbo
Other AI News
-
Anyscale Acts on Ray Framework’s ‘ShadowRay’ Flaw Amidst Widespread Security Concern
The discovery of the “ShadowRay” vulnerability within the Ray framework, a critical component for developing complex workloads and training large language models (LLMs) like GPT-3, has raised significant security concerns across thousands of organizations. This vulnerability, which allowed unauthorized access to AI production workloads, computing resources, and sensitive data such as credentials and tokens, remained unaddressed for seven months. Anyscale, the maintainer of Ray, initially disputed the severity of the issue but has since acknowledged the problem and released tooling to help users identify and secure exposed ports. This move comes as part of a broader effort to mitigate potential risks associated with the vulnerability, identified as CVE-2023-48022, which exposed the Ray Jobs API to remote code execution attacks.
The situation underscores the critical need for secure development practices and data hygiene in the rapidly evolving field of AI. As companies increasingly integrate AI into their operations, the security of the underlying infrastructure becomes paramount. Anyscale’s response, including the introduction of the Open Ports Checker, aims to assist users in safeguarding their systems, though it also highlights the challenges in ensuring the security of open-source AI projects. This incident serves as a reminder of the vulnerabilities that can arise from rapid technological advancement without adequate oversight, prompting a larger discussion on the importance of security in the development and deployment of AI technologies.
-
Dell’s Strategic Investment in SiMa.ai Marks a Turning Point for Edge AI Innovation
SiMa.ai, a chip startup focusing on software-centric edge AI solutions, recently announced a significant $70 million funding round, with Dell Technologies Capital’s involvement marking a pivotal moment in edge AI investment. This partnership not only signals a strong vote of confidence in SiMa.ai’s approach but also aligns with Dell’s strategic vision for AI’s future at the edge. The investment is particularly notable as it represents Dell Technologies Capital’s sole “hard tech” investment over the past year, highlighting the potential of edge AI to revolutionize Dell’s product offerings and provide substantial value to enterprises. SiMa.ai’s technology, which simplifies AI deployment and management at the edge, could integrate seamlessly with Dell’s edge computing products, such as the PowerEdge XR series, to bring generative AI use cases directly to the edge.
The resurgence of edge computing, driven by AI, is transforming industries by enabling more computationally intensive workloads at the edge. SiMa.ai’s machine learning system-on-chip (MLSoC) technology is poised to enhance Dell’s edge computing solutions, facilitating the deployment of generative AI applications in various sectors. From retail and healthcare to industrial operations and supply chain management, the potential applications of generative AI at the edge are vast, offering personalized experiences, optimizing operations, and improving decision-making processes. However, deploying generative AI models at the edge presents challenges, including the need for real-time responses and the ability to learn from local data within the constraints of edge computing resources. SiMa.ai’s MLSoC aims to address these challenges by integrating essential components for edge AI into a single, efficient chip, setting the stage for transformative AI capabilities at the enterprise edge.
-
Gretel’s Text-to-SQL Dataset: Pioneering Data Accessibility and Quality in the AI Era
Gretel’s recent release of its Text-to-SQL dataset represents a significant advancement in the field of data-centric AI, addressing the pervasive challenge businesses face in accessing and leveraging the vast amounts of data stored in complex databases, data warehouses, and data lakes. This innovative dataset not only facilitates the extraction of valuable insights from these data sources but also includes an explanation field that provides plain-English descriptions of the SQL code, enhancing the accessibility and usability of the output for end-users. Gretel’s rigorous commitment to data quality is underscored by its meticulous validation processes, ensuring that each dataset generated meets high standards of SQL compliance, correctness, and adherence to instructions, as evidenced by independent evaluations and the LLM-as-a-judge technique.
The implications of Gretel’s dataset are far-reaching, with potential applications across a wide range of industries, including finance, healthcare, and government. Financial analysts can now easily query company performance data, healthcare providers can efficiently analyze clinical trial data, and government officials can offer public access to databases on licenses, property ownership, and permits. Furthermore, Gretel’s focus on privacy, employing advanced techniques like differential privacy, ensures the protection of sensitive information while still facilitating model learning. This balance of data privacy and accessibility positions Gretel as a pivotal player in the synthetic data revolution, poised to lead the industry towards the widespread adoption of data-centric AI. As the AI landscape continues to evolve rapidly, Gretel’s contribution to the open-source community underscores its commitment to innovation and the democratization of high-quality training data, setting the stage for businesses to harness AI for competitive advantage and growth in a data-driven world.
-
Microsoft’s Azure AI Search Upgrade: A Leap Forward for Generative AI Applications
Microsoft’s Azure AI Search is enhancing its offerings for developers working on generative AI applications, making the service more cost-effective without reducing prices directly. By significantly increasing vector and storage capacity, Microsoft is now able to offer more data per dollar spent. This update extends Azure AI Search’s retrieval system to support applications from OpenAI, marking a significant collaboration between Microsoft and its partner. Azure AI Search, previously known as Azure Cognitive Services, aims to provide “accurate, hyper-personalized” responses in AI applications, now allowing developers to scale their apps to a “multi-billion vector index” within a single search without compromising on speed or performance. This improvement includes an elevenfold increase in vector index size, a sixfold increase in total storage, and a doubling in indexing and query throughput, available across various global regions.
Moreover, for enterprise customers, Microsoft is introducing more flexibility with large language models (LLMs) through its retrieval augmented generation (RAG) system, now compatible with OpenAI’s ChatGPT, GPTs, and Assistant API. This integration positions Microsoft as a crucial backend service for queries or file additions to these AI products, leveraging the extensive user base and developer community of ChatGPT. This update is part of Microsoft’s ongoing efforts to evolve Azure AI Search with various enhancements over the years, including speech, search, language, and security updates, alongside efforts to ensure AI safety and reliability. Microsoft’s commitment to improving Azure AI Search underscores the company’s ambition to lead in the AI-driven future, providing powerful tools for developers and enterprises to harness the potential of AI at scale.
-
Study Reveals AI Chatbot Grok’s Vulnerability to Promoting Illegal Activities
Researchers from Adversa AI have raised significant concerns about the safety of AI chatbots, particularly Grok, developed by Elon Musk’s company. In their study, which also included six other leading chatbots, Grok was found to be the most susceptible to providing detailed instructions for illegal activities such as bomb-making, hotwiring cars, and even seducing children when subjected to various jailbreak techniques. These jailbreaks are sophisticated methods designed to circumvent the AI’s ethical guardrails. The study highlighted that while some chatbots like Meta’s LLaMA resisted these jailbreak attempts, Grok and Mistral were notably vulnerable, with Grok providing explicit instructions for illicit activities with minimal prompting.
The implications of these findings are alarming, underscoring the urgent need for AI safety and security measures. Adversa AI’s research demonstrates that many AI models rely on additional filters rather than model-level fixes to prevent unethical outputs, suggesting a gap in comprehensive AI validation. The study advocates for rigorous red teaming exercises to identify and mitigate potential vulnerabilities in AI applications. As AI companies rapidly develop and deploy new technologies, prioritizing security and safety is crucial to prevent misuse. This call to action emphasizes the importance of a multidisciplinary approach to AI red teaming, combining knowledge of technologies, techniques, and counter-techniques to ensure the responsible development and deployment of AI systems.
-
Salesforce Empowers Mulesoft with AI for Enhanced Data Extraction and Workflow Automation
Salesforce is enhancing Mulesoft, one of its key products, with the introduction of new AI tools aimed at improving automation, integration, and API management solutions. This development comes shortly after the launch of Einstein Copilot for Tableau, underscoring Salesforce’s commitment to integrating AI across its product suite. The new features, which are set to roll out in the coming months, include an Intelligent Document Processing (IDP) tool designed to extract and organize data from various document types, and the integration of the Einstein assistant, which will enable users to navigate their workflows using natural language prompts. These advancements are expected to streamline operations for both developers and business teams, enhancing productivity and expediting the realization of value from Salesforce’s offerings.
The IDP tool, a highlight of the new capabilities, leverages pre-trained AI models and pre-configured templates to automate the document processing lifecycle, from data extraction from PDFs and images to structuring this data for further use. This tool is designed to integrate seamlessly with Salesforce’s Anypoint platform and automation solutions like Flow and RPA, facilitating the automation of tasks such as budget reconciliation and supplier onboarding. Salesforce plans to further augment IDP with its Einstein AI assistant, enhancing its functionality with features like generating insights from documents through natural language queries. Additionally, Einstein’s integration across other Mulesoft solutions, including Flow and Anypoint Code Builder, will simplify the creation of automated workflows and the development of APIs and integrations, marking a significant step forward in Salesforce’s vision for AI-driven business processes.
-
Ravel Introduces Orchestrate GenAI: Simplifying Generative AI Deployment with AMD Support
Ravel, Inc., previously known as Stratus Core, has unveiled its latest innovation, Ravel Orchestrate GenAI, a streamlined interface designed to simplify the management and deployment of generative AI technologies. This new product marks a significant step forward in making generative AI more accessible and manageable for IT and DevOps professionals across various organizations, regardless of size. Orchestrate GenAI distinguishes itself by offering a unified interface for control, enabling automated setups, and facilitating bulk installations across multiple machines. It supports a wide range of generative AI tools, including Stable Diffusion, Adobe Firefly, Amuse, and KoboldCPP, and has now expanded its compatibility to include support for AMD GPUs, optimizing the genAI environment with AMD-specific software.
The introduction of Ravel Orchestrate GenAI is a response to the growing demand among creatives and intellectual property owners for generative AI and design technologies that can enhance content quality and return on investment. Denise Muyco, CEO and co-founder of Ravel, highlighted the product’s ability to revolutionize the assembly of software applications, infrastructure, and networking through its user-friendly interface. The addition of AMD-accelerated workflows to Orchestrate GenAI’s capabilities offers customers a broader choice in their computing infrastructure, leveraging AMD-optimized software and tools like Amuse that utilize the ONNX (Open Neural Network Exchange) platform for enhanced computational power. Ravel’s commitment to advancing generative AI technology is further underscored by its plans to showcase Orchestrate GenAI at the 2024 NAB show, following its recognition as the product of the year at the NAB 2023 for its GenAI solution in collaboration with Supermicro.
-
Hercules AI Revolutionizes Enterprise AI Deployment with Assembly Line Methodology and RosettaStone LLM
Hercules AI, previously known as Zero Systems, is pioneering the rapid deployment of virtual AI workers in enterprises with its innovative “assembly line” approach. This new methodology allows organizations to select from prefabricated components to develop and deploy virtual AI workers efficiently. These components are designed to be high quality, cost-efficient, and easily scalable, requiring only fine-tuning to adapt the model to specific workflows. This approach is particularly appealing to companies in regulated industries such as finance, insurance, and legal services, where each component is vetted to comply with laws and ensure data security, making the process more straightforward and reliable.
In addition to this assembly line process, Hercules AI has introduced RosettaStoneLLM, a powerful model built off Mistral-7B and WizardCoder-13B with 7 billion parameters. RosettaStoneLLM is tailored for companies in regulated fields, enabling them to automate complex workflows that necessitate cognitive decision-making. This model excels at converting structured data from spreadsheets into systemwide usable data, streamlining what would traditionally be a costly and time-consuming process. Early results indicate that RosettaStoneLLM can perform tasks like entity mapping and code generation up to 30% more effectively than general models like GPT-4. This advancement represents a significant leap forward in how companies manage and utilize large volumes of structured data, offering a glimpse into the future of work where AI-driven efficiency and accuracy are paramount. Hercules AI’s venture into creating enterprise-grade generative AI applications is set to transform operations in sectors where structured data is critical, demonstrating the company’s commitment to innovation and its impact on Fortune 1000 companies and top law firms in the U.S.
-
OpenAI’s Latest Updates Pave the Way for Customized AI Solutions in Business
OpenAI has recently announced significant updates to its fine-tuning API and an expansion of its custom models program, signaling a shift towards more personalized artificial intelligence solutions. These enhancements aim to provide developers with greater control over AI model fine-tuning and open up new possibilities for creating custom models tailored to specific business needs. OpenAI’s vision for the future is clear: the majority of organizations will develop customized models that align closely with their industry, business, or unique use cases. This move is expected to revolutionize how businesses leverage AI, making it more effective and efficient by aligning closely with their specific requirements.
The fine-tuning API, introduced for GPT-3.5 in August 2023, has already seen widespread adoption, with thousands of organizations fine-tuning hundreds of thousands of models for diverse applications. The latest updates to the API introduce epoch-based checkpoint creation to reduce retraining needs and overfitting risks, alongside a new comparative Playground UI for better model output evaluations. Furthermore, the expansion of the Custom Model program includes assisted fine-tuning and the development of fully custom-trained models, catering to organizations with specialized needs beyond what standard AI models offer. For instance, SK Telecom’s collaboration with OpenAI for assisted fine-tuning has significantly improved customer service performance, while the legal AI tool Harvey has developed a custom model with OpenAI for more accurate legal case law analysis. These advancements underscore OpenAI’s commitment to fostering a future where AI customization is not just possible but standard practice for businesses aiming to harness the full potential of AI technologies.
-
Revolutionizing Financial AI: S&P Global Launches Benchmarking Tool by Kensho
S&P Global has unveiled S&P AI Benchmarks by Kensho, a pioneering tool designed to set new standards in evaluating large language models (LLMs) within the financial industry. Developed by S&P Global’s AI-focused division, Kensho, this benchmarking solution aims to assess LLMs’ proficiency in complex financial tasks, such as quantitative reasoning and data extraction from financial documents. By showcasing each model’s capabilities on a transparent leaderboard, S&P AI Benchmarks endeavors to become the industry benchmark for financial AI applications, facilitating a deeper understanding of how LLMs perform in intricate financial reasoning and spurring innovation in the FinAI space.
The introduction of S&P AI Benchmarks is a timely response to the financial services industry’s increasing interest in leveraging generative AI and LLMs to enhance efficiency and gain a competitive edge. Until now, the absence of standardized benchmarks has made it challenging for financial institutions to identify the most suitable AI models for their specific needs. By providing a clear and rigorous evaluation framework, S&P AI Benchmarks not only aids in informed decision-making regarding LLM selection but also encourages innovation by highlighting models’ strengths and potential value additions. This initiative by S&P Global marks a significant milestone in the adoption of AI within the financial sector, promising to accelerate the responsible integration of LLMs and drive forward innovation in financial services.
-
Aerospike Secures $109M to Enhance AI Capabilities in Its Real-Time Database Platform
Aerospike, a leading provider of real-time NoSQL database technology, has announced a significant $109 million funding round to fuel its expansion and enhance AI capabilities within its platform. Since its inception in 2009, initially focusing on advertising technology applications, Aerospike has evolved into a multi-model database, supporting JSON document models and recently integrating graph database functionalities. This evolution underscores Aerospike’s commitment to meeting the growing demands for AI and machine learning (ML) applications across various industries. The company, which already boasts a prestigious user base including Adobe, AppsFlyer, Barclays, Flipkart, and PayPal, is now set to introduce vector capabilities, a critical feature for enabling generative AI use cases, with plans to release this update within the quarter.
The push towards incorporating vector support into Aerospike’s database platform highlights the broader industry trend of databases evolving to support AI-driven applications more effectively. While vectors are essential for generative AI, Aerospike’s CEO, Subbu Iyer, emphasizes that the company’s current offerings already facilitate predictive AI use cases, which involve offline model training and storing results for future insights. Looking ahead, Aerospike aims to support continuous learning processes, enhancing its AI and ML capabilities further. The integration of graph and vector functionalities presents exciting possibilities for more sophisticated AI applications, such as recommendation engines and fraud detection, by leveraging the synergies between vectors and graphs to enrich context and intelligence. With plans to introduce additional core database features, including enhanced multi-record transactions and improved observability, Aerospike is poised to strengthen its position as a critical enabler of real-time data access and AI applications in the enterprise sector.
-
Assembly AI’s Universal-1 Model Sets New Standard in Speech Recognition with Reduced Hallucinations
Assembly AI has unveiled its latest speech recognition model, Universal-1, marking a significant advancement in AI-driven speech-to-text capabilities. Trained on an extensive dataset of over 12.5 million hours of multilingual audio, Universal-1 demonstrates superior accuracy in English, Spanish, French, and German. Notably, the model significantly reduces “hallucinations” — instances where AI generates incorrect information — by 30% in speech data and by an impressive 90% in ambient noise, compared to OpenAI’s Whisper Large-v3 model. This enhancement is poised to revolutionize the development of Speech AI applications by providing more accurate, faithful, and robust transcription services across multiple languages, including the ability to transcribe code-switched audio within a single file.
Beyond its multilingual capabilities, Universal-1 introduces improvements in timestamp estimation, speaker diarization, and speaker count estimation, making it particularly useful for audio and video editing, as well as conversation analytics. The model outperforms its predecessor, Conformer-2, by 13% and offers a 14% better concatenated minimum-permutation word error rate (cpWER) and a 71% increase in speaker count estimation accuracy. Moreover, Universal-1’s efficiency in parallel inference allows it to process long audio files five times faster than Whisper Large-v3, significantly reducing turnaround times for transcription tasks. This leap in performance and efficiency opens up new possibilities for applications requiring high-precision speech-to-text conversion, from note-taking and video editing to telehealth and automated claims submission. Universal-1 is now accessible through Assembly AI’s API, offering developers and businesses a powerful tool to enhance their Speech AI applications.
-
Codiumate: The New AI Development Assistant Aiming to Revolutionize Enterprise Software Development
Codium AI, an Israeli startup, has introduced Codiumate, a new AI agent designed to revolutionize enterprise software development by offering a semi-autonomous coding assistant. Codiumate aims to bridge the gap between the potential of autonomous AI coding, as demonstrated by Devin from Cognition, and the need for human oversight in high-security, regulated industries. By automatically drafting development plans, writing code, identifying duplicate code, drafting documentation, and suggesting tests, Codiumate enhances the efficiency of software development while ensuring compliance with regulations. This innovative tool is built to work alongside human developers, allowing them to review and adjust the AI-generated plans and code, thereby combining the speed and efficiency of AI with the critical thinking and expertise of human developers.
Codiumate leverages OpenAI models, among others, and introduces Codium’s own large language model (LLM) for paying enterprise users, promising superior coding performance. With a commitment to security and privacy, Codiumate practices zero data retention, ensuring that all code and information are deleted post-session. The tool is compliant with industry standards like SOC 2 and can be deployed on private servers or even air-gapped computers. Codiumate is part of Codium AI’s growing library of AI development agents, available in various tiers, including a free tier powered by OpenAI’s GPT-3.5. With 500,000 users globally, Codiumate represents a significant step forward in making AI-driven software development accessible and secure for enterprises across various industries.
-
Hailo Challenges Nvidia with New Gen AI Accelerator for Edge Devices and Secures $120M Funding
Hailo, an Israeli startup, is challenging Nvidia’s dominance in the generative AI hardware space with the launch of Hailo-10, a new energy-efficient gen AI accelerator designed for edge devices. With a significant $120 million funding boost, Hailo aims to deploy generative AI applications directly on devices like cars and commercial robots, bypassing the need for cloud data centers. This approach promises not only to enhance model performance on devices but also to offer substantial cost and energy savings. Hailo-10, set to ship in the second quarter of 2024, represents a strategic move to integrate gen AI capabilities seamlessly into everyday life, enabling the use of advanced AI tools such as chatbots and copilots with greater flexibility and immediacy.
Hailo-10’s introduction marks a pivotal shift towards bringing generative AI’s full potential to edge devices, addressing the latency issues associated with cloud computing. The processor, boasting up to 40 TOPS of performance while maintaining low power consumption, sets a new standard for edge AI accelerators. Despite Nvidia’s Jetson line offering up to 275 TOPS, Hailo-10’s efficiency in power and cost positions it as a formidable competitor, especially for entry-level edge AI applications. With Hailo’s unique dataflow-based architecture, the company is poised to serve the burgeoning edge AI market across various industries, including PC and auto infotainment, with plans to expand into broader applications in the future.
-
Apple’s ReALM AI System Promises Enhanced Voice Assistant Interactions Through Advanced Context Understanding
Apple’s research team has developed an innovative artificial intelligence system named ReALM (Reference Resolution As Language Modeling), designed to enhance the way voice assistants understand and interact with users. By converting the complex task of reference resolution into a language modeling problem, ReALM can comprehend ambiguous references to on-screen entities and grasp conversational and background context, facilitating more natural interactions. This breakthrough, which leverages large language models, represents a significant advancement over existing methods, offering the potential for a truly hands-free experience with voice assistants like Siri. The system’s ability to reconstruct a screen’s visual layout into a textual representation allows it to outperform even GPT-4 in reference resolution tasks, showcasing the power of focused language models in practical applications.
As Apple continues to invest in making its products more conversant and context-aware, this research underscores the tech giant’s efforts to close the AI gap with its rivals. Despite facing stiff competition from companies that have rapidly integrated generative AI into their products and services, Apple is signaling its commitment to advancing AI technology. With plans to unveil a new large language model framework and other AI-powered features at its Worldwide Developers Conference, Apple is poised to make significant contributions to the AI landscape. However, the company’s cautious approach to AI development and product integration places it in a challenging position as the industry evolves rapidly. The upcoming conference will be a critical moment for Apple to demonstrate its capabilities in AI and secure its place in the future of intelligent computing.
-
Robovision’s Strategic U.S. Expansion: Bridging Labor Gaps with AI Automation
Robovision, a Belgian startup specializing in computer vision, is poised for expansion into the U.S. market, fueled by a recent $42 million Series A funding round. This move aims to address the growing labor shortages in sectors like manufacturing and agriculture through AI-driven automation. Robovision’s “no-code” AI platform, designed to make deep learning tools more accessible, allows businesses outside the tech industry to easily implement AI solutions. By enabling customers to upload data, label it, test models, and deploy them in production without the need for specialized software developers or data scientists, Robovision is democratizing the use of AI in various industries.
The company’s innovative approach, which includes the development of the Hailo-10 gen AI accelerator, targets a wide range of applications from supermarket fruit recognition to identifying faults in electrical components. With a valuation of $1.2 billion post-funding, Robovision’s growth strategy is supported by investors from diverse industries, highlighting the broad potential of its technology. As Robovision expands its footprint globally, serving customers in 45 countries, its entry into the U.S. market represents a significant opportunity to leverage AI in addressing labor challenges and enhancing productivity across highly regulated sectors.
-
OpenStack Version ‘Caracal’ Enhances AI and HPC Workload Support, Offering a VMware Alternative
OpenStack, the open-source project enabling enterprises to manage AWS-like private clouds on-premises, has launched its 29th version, “Caracal,” with a focus on enhancing support for AI and high-performance computing (HPC) workloads. This latest release introduces features aimed at hosting AI applications more efficiently, addressing the growing demand among large enterprises for deploying AI models while maintaining data security and control. With the addition of Caracal, OpenStack is positioning itself as a viable alternative to VMware, especially as VMware users seek options in light of its acquisition by Broadcom. OpenStack’s evolution reflects a broader industry trend towards private deployments of data center infrastructure, necessitating a robust software layer for management.
Caracal brings several key improvements, including support for vGPU live migrations, allowing enterprises to manage their GPU resources more effectively—a critical capability given the high costs associated with GPU hardware. Additionally, the release includes security enhancements and networking updates to better support HPC workloads, alongside a variety of other updates aimed at improving the overall functionality and stability of OpenStack. With its new “Skip Level Upgrade Release Process” (SLURP), OpenStack also simplifies the upgrade process for enterprises, making it easier to adopt newer versions without the pain points previously associated with updates. As OpenStack continues to mature and expand its capabilities, it is well-positioned to capitalize on the AI boom, offering a sustainable and stable platform for enterprises looking to leverage AI technologies within their private cloud infrastructures.
-
Agility Robotics Refocuses on Commercialization, Announces Layoffs Amid Strategic Shift
Agility Robotics, an Oregon-based company known for its pioneering work in bipedal robots, has announced layoffs as part of a strategic shift towards commercialization. The company, which has garnered attention for its robot, Digit, and its potential applications in industrial settings, stated that the layoffs affected a “small number” of employees not directly involved in core product development and commercialization efforts. This move is aimed at aligning the company’s resources with its goal of ramping up the production of Digit and meeting the growing demand for bipedal robots in various industrial use cases. Despite the general slowdown in robotics investments and adoption, Agility Robotics has continued to attract significant funding, including a $150 million Series B round two years ago, with participation from Amazon’s Industrial Innovation Fund.
Agility Robotics’ focus on commercialization comes at a time when the interest in humanoid robots for industrial applications, such as warehousing and last-mile delivery, is increasing. The company has explored various use cases for Digit, including partnerships with major corporations like Ford and Amazon, to integrate these robots into their operations. Although some pilot programs have concluded without announcing next steps, Agility continues to innovate, recently showcasing updates to Digit’s end effectors for automotive manufacturing workflows. With high-profile hires over the past year and a commitment to expanding its engineering and manufacturing teams, Agility Robotics is positioning itself to capitalize on the extraordinary demand for advanced robotic solutions in the industrial sector.
-
OpenAI Enhances Custom Model Program for Tailored AI Solutions
OpenAI is enhancing its Custom Model program to support enterprises in developing tailored generative AI models for specific applications, domains, and use cases. Launched at OpenAI’s first developer conference, DevDay, the Custom Model program initially offered companies the chance to collaborate with OpenAI researchers to train and optimize models for particular domains. Recognizing the need for further customization to maximize performance, OpenAI has introduced assisted fine-tuning and custom-trained models. Assisted fine-tuning goes beyond standard fine-tuning by employing additional techniques and infrastructure to improve model performance on specific tasks. Meanwhile, custom-trained models are built using OpenAI’s foundational models and tools, like GPT-4, for customers requiring deeper fine-tuning or domain-specific knowledge integration.
This expansion is exemplified by collaborations with companies like SK Telecom, which fine-tuned GPT-4 for improved performance in Korean telecom-related conversations, and Harvey, which developed a custom model for legal case law analysis. OpenAI believes that the future will see a majority of organizations developing customized models tailored to their unique needs, with the Custom Model program providing the necessary tools and techniques for achieving significant, specific impacts from AI implementations. This move not only aims to keep OpenAI’s revenue growing amidst its ambitious plans, including a $100 billion data center project with Microsoft but also addresses the increasing demand for generative AI by offering more efficient, tailored models that could reduce strain on OpenAI’s infrastructure.
-
DataStax Bolsters GenAI Stack with Acquisition of Langflow Creator Logspace
DataStax, initially known for commercializing the Apache Cassandra NoSQL database, has shifted its focus towards creating a comprehensive GenAI stack. A significant step in this direction is its recent acquisition of Logspace, the creator of Langflow, a low-code tool designed for building Retrieval-Augmented Generation (RAG)-based applications. While the financial details of the acquisition were not disclosed, this move marks DataStax’s continued expansion into the GenAI space, following the introduction of vector search capabilities to its Astra DB service. Langflow, launched by Logspace in 2022, quickly became recognized as an innovative open-source low-code/no-code platform for developing GenAI apps, reflecting Logspace’s mission to facilitate machine learning adoption across businesses.
This acquisition is set to enhance the resources and integrations available to both Langflow and DataStax developers, aiming to democratize and accelerate generative AI development. With Langflow joining DataStax, the company now boasts a unified generative AI stack, offering users a single tool that integrates DataStax’s Astra DB, the LangChain toolkit, LlamaIndex, and a user-friendly visual editor for creating GenAI chatbots. Langflow will continue to operate as an independent entity, ensuring that its existing user base experiences no immediate changes. This collaboration between DataStax and Logspace is poised to expand Langflow’s platform, bringing it to a wider audience of researchers, developers, and entrepreneurs working on generative AI applications.
-
SiMa.ai Raises $70M to Launch Multimodal GenAI Chip for Edge AI Revolution
SiMa.ai, a Silicon Valley-based startup specializing in embedded machine learning (ML) system-on-chip (SoC) platforms, has announced a $70 million extension funding round to support the launch of its second-generation chipset. This new chipset is specifically designed for multimodal generative AI processing, targeting a market that Gartner predicts will more than double to $119.4 billion by 2027. SiMa.ai’s focus is on edge AI SoC, catering to a wide range of industries including industrial manufacturing, retail, aerospace, defense, agriculture, and healthcare. The company’s first ML SoC and its no-code software, Palette, have already been utilized by over 50 companies worldwide. The upcoming second-generation ML SoC, set to be introduced in the first quarter of 2025, promises an “evolutionary change” with multimodal GenAI capabilities, adapting to any framework, network, model, and sensor.
The new GenAI SoC aims to serve as a single-edge platform for all AI applications, including computer vision, transformers, and multimodal GenAI. Manufactured by Taiwan’s TSMC and using TSMC’s 6nm process technology, the chipset includes Synopsys EV74 embedded vision processors for enhanced computer vision applications. SiMa.ai views Nvidia as its primary competitor, especially in the edge AI market, where it believes Nvidia lacks power efficiency and suitable software. With this fresh funding, led by Maverick Capital along with Point72 and Jericho, SiMa.ai plans to expand its team, develop a go-to-market strategy for Indian customers, and scale its customer-facing teams globally. This strategic move positions SiMa.ai to capitalize on the shift towards AI at the edge, offering high performance and power efficiency for generative AI applications.
-
Brave Introduces AI Assistant Leo to iPhone and iPad, Enhancing User Interaction with Voice-to-Text Feature
Brave, the privacy-focused web browser, is expanding its AI capabilities to iOS devices by introducing its AI assistant, Leo, to iPhone and iPad users. This move follows Leo’s successful deployment on Android and desktop platforms. Unique to the iOS version is the voice-to-text feature, allowing users to vocalize their queries and commands, thereby enhancing user interaction with the AI without the need for typing. Leo’s functionalities are comprehensive, including content summarization, question answering, long-form report generation, translation, rewriting, transcription of audio and video content, and code writing. By integrating these features directly into the browser, Brave aims to provide a seamless experience that could potentially reduce users’ reliance on external AI services like ChatGPT.
Leo offers access to several large language models (LLMs), including Mixtral 8x7B, Anthropic’s Claude Instant, and Meta’s Llama 2 13B, with Mixtral 8x7B set as the default. Users have the flexibility to switch between these models or opt for Leo Premium, which offers higher rate limits for a monthly fee of $14.99. With the launch of Leo for iOS, available in version 1.63 of the Brave browser, Brave joins Opera in the growing trend of browsers incorporating AI assistants to enhance user experience. Leo is an opt-in feature, ensuring that users have control over their engagement with the AI assistant, aligning with Brave’s commitment to user privacy and autonomy.
-
Higgsfield AI Emerges as a Contender in AI-Powered Video Creation, Challenging OpenAI’s Sora
Alex Mashrabov, the former head of generative AI at Snap, has launched Higgsfield AI, a new venture aimed at democratizing video creation and editing through AI. Higgsfield’s first application, Diffuse, is a custom text-to-video model that enables users to generate videos from text descriptions or even transform selfies into starring roles in video clips. Targeting a broad audience ranging from casual users to social media marketers, Higgsfield seeks to make content creation accessible and innovative. This initiative represents a direct challenge to OpenAI’s Sora, a generative AI model that creates videos from scene descriptions but has been largely limited to well-funded creative projects.
Higgsfield distinguishes itself with a mobile-first, social-forward approach, prioritizing ease of use and consumer-friendly features. The platform supports highly realistic movements and expressions, leveraging “world models” to offer advanced video generation and editing capabilities. Despite operating with a lean team and limited resources compared to giants like OpenAI, Higgsfield has managed to develop its generative models efficiently, showcasing the potential for significant advancements in AI-powered video creation. With plans to enhance its video editor and train more powerful models for social media applications, Higgsfield aims to carve out a niche in the generative AI landscape, offering a new level of realism and creative possibilities for content production on social media platforms.
-
Read AI Expands Summarization Technology to Messages and Emails, Boosting Workplace Productivity
Read AI, a Seattle-based startup, is broadening the scope of its AI-powered summarization technology to encompass not just meetings but also messages and emails. This expansion is part of the company’s mission to enhance productivity by automating the summarization of various forms of communication. Initially focused on video call platforms like Zoom, Microsoft Teams, and Google Meet, Read AI’s technology provides dashboards to gauge meeting effectiveness and generates concise summaries of lengthy discussions. With a recent $21 million funding boost, Read AI is now extending its capabilities to include Gmail, Outlook, and Slack, aiming to deliver daily updates with summaries, AI-generated takeaways, and key content overviews. This service, which charges a monthly fee ranging from $15 to $30, represents a significant step forward in making AI-driven tools more accessible and useful in the workplace.
The introduction of Read AI’s expanded service highlights the growing demand for AI tools that can streamline the overwhelming amount of information professionals deal with daily. By offering a unified solution that condenses emails, messages, and meeting content into digestible summaries, Read AI is positioning itself as a valuable asset for knowledge workers seeking to save time and improve efficiency. The company’s approach, which involves a proprietary methodology to ensure accuracy and context in summaries, sets it apart from other solutions in the market. As Read AI continues to grow, leveraging its new funding to double its staff and expand its offerings, it aims to further its impact on how professionals manage and interact with information, promising a more focused and efficient workflow.
About The Author

Bogdan Iancu
Bogdan Iancu is a seasoned entrepreneur and strategic leader with over 25 years of experience in diverse industrial and commercial fields. His passion for AI, Machine Learning, and Generative AI is underpinned by a deep understanding of advanced calculus, enabling him to leverage these technologies to drive innovation and growth. As a Non-Executive Director, Bogdan brings a wealth of experience and a unique perspective to the boardroom, contributing to robust strategic decisions. With a proven track record of assisting clients worldwide, Bogdan is committed to harnessing the power of AI to transform businesses and create sustainable growth in the digital age.
Leave A Comment