Running out of memory while processing large responses - DSS
Hi All,
I believe I am not using best practices when processing the response. for ex., chunking, paging, or streaming. Hence my application goes out of memory (breaching 500MB). I would like to get your advice on how to avoid an oom situation by processing the data efficiently.
I tried looking at the examples on pages: https://selectapi.datascope.refinitiv.com/RestApi.Help/Home/KeyMechanisms?ctx=Extractions&tab=0&uid=StreamingJson and https://developers.lseg.com/en/api-catalog/datascope-select/datascope-select-rest-api/download, however, I did not find any examples I can implement for my use case.
Here are the steps in my app using Java 17:
I call the following end point with 6000 instruments to retrieve 6 data points back.
https://selectapi.datascope.refinitiv.com/RestApi/v1/Extractions/ExtractWithNotes
"@odata.type": "#DataScope.Select.Api.Extractions.ExtractionRequests.EndOfDayPricingExtractionRequest"
This request times out as expected and when I poll using locationUrl, I end up getting the data I need.
https://selectapi.datascope.refinitiv.com/RestApi/v1/Extractions/ExtractRawResult(ExtractionId='test')
I store this data in a list. Then, I call the following end point with the same 6000 instruments to retrieve 1 data point back. Reason being, this particular data point is only available in this API.
https://selectapi.datascope.refinitiv.com/RestApi/v1/Extractions/ExtractWithNotes
"@odata.type": "#DataScope.Select.Api.Extractions.ExtractionRequests.FixedIncomeAnalyticsExtractionRequest",
I store this data in a separate list. Then using a hashmap and for each loop, I merge the requests into a single list of objects, with each object representing data for a unique instrument identified by its instrumentISIN.
Best Answer
-
If the application keeps the data of all 6000 instrumetns in the application's memory. I am not sure how paging, chunking or streaming can reduce the size of memory used to store the data of 6000 instruments. The total size of data will be the same for 6000 instruments.
You may reduce the number of instruments in each request, such 3000 instruments in each request. Otherwise, you may save the extraction resutls in the local files and process those files according to the application's resouces and requirements.
0
Answers
-
Thank you for reaching out to us.
Typically, you can increase the size of memory by using the JVM options.
-Xms<size> set initial Java heap size
-Xmx<size> set maximum Java heap size
-Xss<size> set java thread stack sizeYou can check this discussion on StackOverflow for more information.
To verify the memory usage in the application, you can use the Java profiler tool, such as VisualVM to identify which components in the application consume a lot of memory.
0 -
Thanks, however, I am not looking to expand the memory.
Do you know if I can use paging, chunking or streaming to make the processing more efficient?
0
Categories
- All Categories
- 6 AHS
- 37 Alpha
- 161 App Studio
- 4 Block Chain
- 4 Bot Platform
- 16 Connected Risk APIs
- 47 Data Fusion
- 30 Data Model Discovery
- 608 Datastream
- 1.3K DSS
- 577 Eikon COM
- 4.9K Eikon Data APIs
- 7 Electronic Trading
- Generic FIX
- 7 Local Bank Node API
- Trading API
- 2.7K Elektron
- 1.3K EMA
- 236 ETA
- 519 WebSocket API
- 33 FX Venues
- 10 FX Market Data
- 1 FX Post Trade
- 1 FX Trading - Matching
- 12 FX Trading – RFQ Maker
- 5 Intelligent Tagging
- 2 Legal One
- 20 Messenger Bot
- 2 Messenger Side by Side
- 9 ONESOURCE
- 7 Indirect Tax
- 59 Open Calais
- 264 Open PermID
- 39 Entity Search
- 2 Org ID
- PAM
- PAM - Logging
- 8.4K Private Comments
- 6 Product Insight
- Project Tracking
- ProView
- ProView Internal
- 20 RDMS
- 1.4K Refinitiv Data Platform
- 367 Refinitiv Data Platform Libraries
- 3 Refinitiv Due Diligence
- LSEG Due Diligence Portal API
- 3 Refinitiv Due Dilligence Centre
- Rose's Space
- 1.1K Screening
- 18 Qual-ID API
- 13 Screening Deployed
- 23 Screening Online
- 10 World-Check Customer Risk Screener
- 990 World-Check One
- 44 World-Check One Zero Footprint
- 45 Side by Side Integration API
- Test Space
- 3 Thomson One Smart
- 1.2K TR Internal
- Global Hackathon 2015
- 2 Specialists Who Code
- 10 TR Knowledge Graph
- 150 Transactions
- 142 REDI API
- 1.7K TREP APIs
- 4 CAT
- 21 DACS Station
- 117 Open DACS
- 1.1K RFA
- 103 UPA
- 172 TREP Infrastructure
- 224 TRKD
- 886 TRTH
- 5 Velocity Analytics
- 5 Wealth Management Web Services
- 59 Workspace SDK
- 9 Element Framework
- 5 Grid
- 13 World-Check Data File
- Yield Book Analytics
- 46 中文论坛