tech:

taffy

OpenAI launches GPT-4 Turbo; unveils API improvements with price reductions

OpenAI has announced the launch of GPT-4 Turbo, an enhancement of the previously released GPT-4 model. The upgraded model boasts a 128K context window, which allows for a larger text prompt capacity, equivalent to over 300 pages of content.

OpenAI has also reduced the cost for both input and output tokens by 3x and 2x respectively, compared to its predecessor.

Developers can now test GPT-4 Turbo by utilizing the gpt-4-1106-preview in the API, with a stable version expected to release soon. Improvements in function calling are also part of the package, allowing for complex tasks like executing multiple commands within a single message.

GPT-4 Turbo has been designed to better understand and execute instructions, a feature complemented by the newly introduced JSON mode, says OpenAI. The mode guarantees JSON output, catering to the needs of developers working with Chat Completions API outside of function calling.

OpenAI is also introducing a ‘seed’ parameter for generating reproducible outputs, useful for debugging and testing. The addition of log probabilities for output tokens will assist developers in creating more interactive features like autocomplete in search functionalities.

Alongside GPT-4 Turbo, OpenAI has upgraded GPT-3.5 Turbo, which now supports a 16K context window and shows a 38% improvement in format following tasks. The new version can be accessed via the gpt-3.5-turbo-1106 API call, and will automatically replace the older model on December 11 for existing applications.

The Assistants API is a new release aimed at simplifying the creation of AI-powered applications. It encompasses the Code Interpreter and Retrieval features, as well as function calling to streamline app development.

GPT-4 Turbo now includes vision capabilities within the Chat Completions API, which can be accessed via gpt-4-vision-preview. This allows for image input, expanding its utility in real-world scenarios.

Developers can also integrate DALL·E 3 into their applications through the Images API. Meanwhile, the Text-to-Speech API now offers human-like speech synthesis with various voice options.

On the customization front, OpenAI is offering experimental access to GPT-4 fine-tuning and launching a Custom Models program for extensive model training.

Developers can begin exploring these new features and improvements, with beta access to the Assistants API now open to all.

You can find the OpenAI blog post with the announcements here.

[Image courtesy: OpenAI]

Just in

Tembo raises $14M

Cincinnati, Ohio-based Tembo, a Postgres managed service provider, has raised $14 million in a Series A funding round.

Raspberry Pi is now a public company — TC

Raspberry Pi priced its IPO on the London Stock Exchange on Tuesday morning at £2.80 per share, valuing it at £542 million, or $690 million at today’s exchange rate, writes Romain Dillet. 

AlphaSense raises $650M

AlphaSense, a market intelligence and search platform, has raised $650 million in funding, co-led by Viking Global Investors and BDT & MSD Partners.

Elon Musk’s xAI raises $6B to take on OpenAI — VentureBeat

Confirming reports from April, the series B investment comes from the participation of multiple known venture capital firms and investors, including Valor Equity Partners, Vy Capital, Andreessen Horowitz (A16z), Sequoia Capital, Fidelity Management & Research Company, Prince Alwaleed Bin Talal and Kingdom Holding, writes Shubham Sharma. 

Capgemini partners with DARPA to explore quantum computing for carbon capture

Capgemini Government Solutions has launched a new initiative with the Defense Advanced Research Projects Agency (DARPA) to investigate quantum computing's potential in carbon capture.