<span class="vcard">/u/Successful-Western27</span>
/u/Successful-Western27

Can GPT models be financial analysts? ChatGPT, GPT-4 fail CFA exams in new study by JP Morgan, Queens University, and Virginia Tech

Researchers evaluated ChatGPT and GPT-4 on mock CFA exam questions to see if they could pass the real tests. The CFA exams rigorously test practical finance knowledge and are known for being quite difficult. They tested the models in zero-shot, few-sho…

Researchers propose GameGPT: A multi-agent approach to fully automated game development

Game dev is super complex nowadays – games have huge codebases, massive teams, and dev cycles dragging on for years. Costs are insane too – budgets can hit $100M+ easily. In a new paper, researchers propose to reverse this trend with an AI framework ca…

Lemur: Harmonizing Natural Language and Code for Language Agents

Today's conversational bots like Claude and GPT can chat impressively but aren't great at complex planning or executing technical tasks. To overcome this, new research from HKU builds open-source AI agents that blend natural language and coding…

Inverting Transformers Significantly Improves Time Series Forecasting

Transformers are great at NLP and computer vision tasks, but I was surprised to learn they still lag behind simple linear models at time series forecasting. The issue is how most Transformer architectures treat each timestamp as a token and fuse all th…

ALMT: Using text to narrow focus in multimodal sentiment analysis improves performance

Multimodal sentiment analysis combines text, audio and video to understand human emotions. But extra inputs can add irrelevant or conflicting signals. So filtering matters. Researchers made a "Adaptive Language-guided Multimodal Transformer" …

Researchers showcase method for AI-based roundabout design to help developing countries improve roadways

I like Cities: Skylines, but struggle at building roundabouts. Turns out, despite being safer than intersections, they're also tricky to design in real life – small tweaks can ruin traffic flow. They're designed iteratively. This is a pain for …

Brown University Paper: Low-Resource Languages (Zulu, Scots Gaelic, Hmong, Guarani) Can Easily Jailbreak LLMs

Researchers from Brown University presented a new study supporting that translating unsafe prompts into `low-resource languages` allows them to easily bypass safety measures in LLMs. By converting English inputs like "how to steal without getting …

DeepMind, Univ. of Illinois: Is self-correction a viable method to improve LLM reasoning? Probably not.

Can LLMs actually improve their own reasoning by self-correcting mistakes? A new paper from DeepMind and the University of Illinois looks to answer this quantitatively. The results show that unaided, LLMs struggle at self-correction for reasoning tasks…

New Paper: Enabling Language Models to Implicitly Learn Self-Improvement From Data

LLMs keep getting more capable at generating natural language. But there's always room for improving the quality and alignment of their responses. Typically this requires lots of human effort to collect more training data. So researchers are explor…

Infinite context windows? Streaming LLMs can be extended to infinite sequence lengths without any fine-tuning.

LLMs like GPT-3 struggle in streaming uses like chatbots because their performance tanks on long texts exceeding their training length. I checked out a new paper investigating why windowed attention fails for this. By visualizing the attention maps, th…