Skip to content

Conversation

sabrenner
Copy link
Collaborator

@sabrenner sabrenner commented Jul 16, 2025

What does this PR do?

Moves streamed response coalescing to the tracing plugin instead of the instrumentation. This allows us:

  1. To start decoupling business logic from the instrumentation/simplifying it in general
  2. Makes it easier to add different streamed implementations for different endpoints going forward (as we plan OpenAI responses soon)

Motivation

Easier maintenance, developer experience, decoupling.

There's still a lot of cleanup needed for this integration that has unfortunately just been lower priority tech debt, but this is hopefully a good start. I figure a good amount of refactoring will come when we migrate this integration to use orchestrion instead.

MLOB-3331

Copy link

Overall package size

Self size: 9.68 MB
Deduped: 109.26 MB
No deduping: 109.65 MB

Dependency sizes | name | version | self size | total size | |------|---------|-----------|------------| | @datadog/libdatadog | 0.7.0 | 35.02 MB | 35.02 MB | | @datadog/native-appsec | 10.0.0 | 20.3 MB | 20.31 MB | | @datadog/native-iast-taint-tracking | 4.0.0 | 11.72 MB | 11.73 MB | | @datadog/pprof | 5.9.0 | 9.77 MB | 10.14 MB | | @opentelemetry/core | 1.30.1 | 908.66 kB | 7.16 MB | | protobufjs | 7.5.3 | 2.95 MB | 5.6 MB | | @datadog/wasm-js-rewriter | 4.0.1 | 2.85 MB | 3.58 MB | | @datadog/native-metrics | 3.1.1 | 1.02 MB | 1.43 MB | | @opentelemetry/api | 1.8.0 | 1.21 MB | 1.21 MB | | jsonpath-plus | 10.3.0 | 617.18 kB | 1.08 MB | | import-in-the-middle | 1.14.2 | 122.36 kB | 850.93 kB | | lru-cache | 10.4.3 | 804.3 kB | 804.3 kB | | source-map | 0.7.4 | 226 kB | 226 kB | | opentracing | 0.14.7 | 194.81 kB | 194.81 kB | | pprof-format | 2.1.0 | 111.69 kB | 111.69 kB | | @datadog/sketches-js | 2.1.1 | 109.9 kB | 109.9 kB | | lodash.sortby | 4.7.0 | 75.76 kB | 75.76 kB | | ignore | 7.0.5 | 63.38 kB | 63.38 kB | | istanbul-lib-coverage | 3.2.2 | 34.37 kB | 34.37 kB | | rfdc | 1.4.1 | 27.15 kB | 27.15 kB | | @isaacs/ttlcache | 1.4.1 | 25.2 kB | 25.2 kB | | dc-polyfill | 0.1.9 | 25.11 kB | 25.11 kB | | tlhunter-sorted-set | 0.1.0 | 24.94 kB | 24.94 kB | | shell-quote | 1.8.3 | 23.74 kB | 23.74 kB | | limiter | 1.1.5 | 23.17 kB | 23.17 kB | | retry | 0.13.1 | 18.85 kB | 18.85 kB | | semifies | 1.0.0 | 15.84 kB | 15.84 kB | | jest-docblock | 29.7.0 | 8.99 kB | 12.76 kB | | crypto-randomuuid | 1.0.0 | 11.18 kB | 11.18 kB | | ttl-set | 1.0.0 | 4.61 kB | 9.69 kB | | mutexify | 1.4.0 | 5.71 kB | 8.74 kB | | path-to-regexp | 0.1.12 | 6.6 kB | 6.6 kB | | koalas | 1.0.2 | 6.47 kB | 6.47 kB | | module-details-from-path | 1.0.4 | 3.96 kB | 3.96 kB |

🤖 This report was automatically generated by heaviest-objects-in-the-universe

Copy link

codecov bot commented Jul 16, 2025

Codecov Report

All modified and coverable lines are covered by tests ✅

Project coverage is 82.79%. Comparing base (e2dbaa8) to head (93021ce).
Report is 11 commits behind head on master.

Additional details and impacted files
@@            Coverage Diff             @@
##           master    #6107      +/-   ##
==========================================
+ Coverage   81.98%   82.79%   +0.80%     
==========================================
  Files         475      475              
  Lines       19598    19600       +2     
==========================================
+ Hits        16067    16227     +160     
+ Misses       3531     3373     -158     

☔ View full report in Codecov by Sentry.
📢 Have feedback on the report? Share it here.

🚀 New features to boost your workflow:
  • ❄️ Test Analytics: Detect flaky tests, report on failures, and find test suite problems.
  • 📦 JS Bundle Analysis: Save yourself from yourself by tracking and limiting bundle sizes in JS merges.

@pr-commenter
Copy link

pr-commenter bot commented Jul 16, 2025

Benchmarks

Benchmark execution time: 2025-07-16 12:53:01

Comparing candidate commit 93021ce in PR branch sabrenner/refactor-openai-streamed-response-handling with baseline commit e2dbaa8 in branch master.

Found 0 performance improvements and 0 performance regressions! Performance is the same for 1271 metrics, 52 unstable metrics.

@sabrenner sabrenner marked this pull request as ready for review July 16, 2025 12:59
@sabrenner sabrenner requested a review from a team as a code owner July 16, 2025 12:59
@sabrenner sabrenner merged commit ab554a7 into master Jul 17, 2025
773 of 774 checks passed
@sabrenner sabrenner deleted the sabrenner/refactor-openai-streamed-response-handling branch July 17, 2025 15:46
Copy link
Collaborator

@BridgeAR BridgeAR left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I was too late as it seems :D

* @param {number} n the number of choices to expect in the response
* @returns {Record<string, any>}
*/
function constructChatCompletionResponseFromStreamedChunks (chunks, n) {
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

We could consolidate these two methods and use a smaller part for the unique handling and just pass through that smaller method. That prevents code duplication :)

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

i thought about that but just wanted to be explicit in how we're handling each case - if that's all good with you! i thought trying to extract it to pass something like a onIndexedChunk function to the shared handler function would hinder readability/initial understanding a bit, but if you feel strongly i can implement it!

ghost pushed a commit that referenced this pull request Jul 18, 2025
@ghost ghost mentioned this pull request Jul 18, 2025
watson pushed a commit that referenced this pull request Jul 21, 2025
tlhunter pushed a commit that referenced this pull request Aug 22, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants