Skip to main content
Filter by
Sorted by
Tagged with
1 vote
0 answers
107 views

I'm trying to fill a certain number of samples from a cpal microphone source (about 10 seconds worth). I want to process those samples as they come in with low latency and in regularly sized blocks of ...
Jason Kleban's user avatar
1 vote
0 answers
62 views

I'm building a RAG-based document QA system using Python (no LangChain), LLaMA (50K context), PostgreSQL with pgvector, and Docling for parsing. Users can upload up to 10 large documents (300+ pages ...
Anton Lee's user avatar
3 votes
1 answer
153 views

I'm trying to read through multiple compressed tables that are 5GB+ in size in R, and because I have insufficient memory to read them into memory all at once I need to process them one chunk at a time,...
Alex Petty's user avatar
0 votes
0 answers
39 views

This is ASP.NET API function to process real-time chatbot with Qwen2.5 local setup with FastAPI. This is current code, but StreamResponse doesn't work correctly. How to optimize this code? [Authorize] ...
MONATE's user avatar
  • 23
0 votes
0 answers
93 views

I am trying to develop a map generation system in Unity. I want to have a system for chunks and have simplified the problem down to a small amount of code, which is not attempting a great deal. I just ...
Seb's user avatar
  • 11
0 votes
0 answers
53 views

I am developing a relatively large model using xarray and therefore want to make use of chunks. Most of my operations run a lot faster when chunked but there is one that keeps running (a lot) slower ...
Rogier Westerhoff's user avatar
0 votes
1 answer
170 views

I am trying Spring Batch with remote partitioning [ master-salve approach]. I have one master step which sends records to worker nodes via KAFKA. All was working fine until parallel job executions ...
Mahantesh Masali's user avatar
0 votes
1 answer
287 views

according to this answer, a proper chunk size is important for optimizing I/O perfromance. I am 3000 jpg images, whose size vary from 180kB to 220kB. I am going to save them as bytes. I know 2 methods ...
zheyuanWang's user avatar
  • 1,462
0 votes
2 answers
375 views

Im doing a RAG model with some documents. Testing Llamaindex SubDocSummaryPack, seems to be a good choice for documents chunking instead of simple chunking the original information. After using ...
Diego's user avatar
  • 11
2 votes
0 answers
183 views

I'm using .NET 8, System.Text.Json 8 HttpClient to make request, HttpConnection.ChunkedEncodingReadStream as input stream from the response, and JsonSerializer.DeserializeAsyncEnumerable to ...
Oleh Hrechukh's user avatar
1 vote
0 answers
465 views

Has anyone tried using Langchain's AI21 integration AI21SemanticTextSplitter? There is a mention of it on Langchain's Text Splitters Page. This is its documentation. I tried the examples given there ...
Akshitha Rao's user avatar
-1 votes
1 answer
2k views

I have this Langchain code for my own dataset: from langchain_community.vectorstores import FAISS from langchain_openai import ChatOpenAI, OpenAIEmbeddings vectorstore = FAISS.from_texts( docs, ...
user17811469's user avatar
1 vote
2 answers
7k views

I have structured data (CSV) that has a column of semantically rich text of variable length. I could mine the data so the CSV file has a max length per row of data by using an LLM to summarize ...
Lance Kind's user avatar
  • 1,116
0 votes
1 answer
420 views

I try to read and process the 8gb csv file using cudf. Reading all file at once doesn't fit neither into GPU memory nor into my RAM. That's why I use the dask_cudf library. Here is the code: import ...
shda's user avatar
  • 734
0 votes
0 answers
123 views

I am trying to implement spring batch remote chunking for heavy file . What should be ideal setup for deployment of remote chunking physically in kubernetes . 1)Can we setup worker and manager in same ...
richa kumari's user avatar
0 votes
1 answer
2k views

I am working on a RAG system using LlamaIndex. I try to adapt small-to-big chunking strategy for retrieval stage. I have numerous articles as inputs and some metadata about them. here is the list of ...
aearslan's user avatar
  • 176
1 vote
0 answers
393 views

I'm using gRPC to send messages from Java to a Python service. Recently I started getting larger messages that sometimes exceed the maximum message size for gRPC. Increasing that size is not possible ...
Noam_I's user avatar
  • 17
-1 votes
1 answer
307 views

I'm working on a ChatGPT integration in Node/Express and would like to first respond to my client with some metadata in JSON, then start streaming ChatGPT's response as it streams in. Currently, there ...
E-Madd's user avatar
  • 4,592
0 votes
0 answers
322 views

So, recently I came across adobe pdf extraction API, I'm using python and for those who aren't aware of adobe's extraction methods, given a PDF the API returns back the extracted text with each ...
Daniel's user avatar
  • 11
0 votes
2 answers
2k views

I have a large database of documents (these “documents” are essentially web pages and they are all in HTML). They have information regarding the business itself and can contain a lot of similar ...
Bruno Vaz's user avatar
0 votes
1 answer
245 views

Working with large datasets in Python via Pandas and initially chunked the two datasets so they could load into memory but not sure how to merge them given they are turned into TextFileReader instead ...
guanhelluh's user avatar
0 votes
1 answer
499 views

can someone please help me to debug this function const chunkify = async (file: Blob) => { const totalSize = file.size; const chunkSize = 1024 * 1024 * 100; // 100MB const chunks = [] as ...
Oyedeji's user avatar
  • 112
2 votes
1 answer
375 views

I am trying to read large csv file (84GB) in chunks with pandas, filter out necessary rows and convert it to df import pandas as pd chunk_size = 1000000 # Number of rows to read per chunk my_df = pd....
Yerassyl Pirzhanov's user avatar
3 votes
1 answer
2k views

Following documentation, and the code snippet provided from https://docs.sqlalchemy.org/en/14/core/connections.html#streaming-with-a-fixed-buffer-via-yield-per (posted directly below), my query is not ...
scrollout's user avatar
  • 594
0 votes
1 answer
212 views

`i tried writing a chunk upload code with js and php, it works fine when you upload a video file less than a 10mb but uploading a video file 17.6mb, the file uploads with it maintains it appropriate ...
Theophilus Tetteh's user avatar
0 votes
1 answer
2k views

I am trying to download a large file >100MB from a gRPC server using gRPC bidirectional streaming. I need to break the file into chunks on the server and stream the bytes. I am not sure how to ...
aytida's user avatar
  • 1
1 vote
0 answers
186 views

A common task in my daily data wrangling is converting tab-delimited text files to xarray datasets and continuing analysis on the dataset and saving to zarr or netCDF format. I have developed a data ...
officialankan's user avatar
0 votes
1 answer
405 views

I have implemented Spring Batch Remote Chunking with Kafka. I have implemented both Manager and worker configuration. I want to send some DTO or object in chunkresponse from worker side to Manager and ...
akashsharma3030's user avatar
0 votes
0 answers
252 views

Cant figure out how to send arrays of objects Currently I have a back-end MongoDB database with multiple collections which I have to query for matching data and send the data back to the client as ...
Muhammad Yahya Warraich's user avatar
1 vote
1 answer
2k views

My original configuration in vue.config.js using the default chunking strategy, which takes about 5 minutes to build locally and 35 minutes in gitlab pipeline, and results in one chunk being > 50MB ...
robbyc73's user avatar
0 votes
1 answer
350 views

I'm a total noob so sorry if I'm asking something obvious. My question is twofold, or rather it's two questions in the same topic: I'm studying nltk in Uni, and we're doing chunks. In the grammar I ...
sickboy83's user avatar
6 votes
0 answers
538 views

I have an artisan command that is scheduler to run every 15 minutes to update a model calculated field. The command process thousands of rows and ends up eating all of the memory. My understanding of ...
naghal's user avatar
  • 726
1 vote
0 answers
896 views

I'm trying to read a ~2.3 GB file that exceeds the RAM of my hardware. I want to apply chunking to read this file. I tried chunksize of 10**2, 10**3, 10**4, and 10**5 but all still exceed the RAM ...
melolilili's user avatar
3 votes
1 answer
3k views

I already read some article but I am still confused. I n pagination it will execute a query when load a page but what happen in chunk ? I read https://laravel-school.com/posts/laravel-pagination-vs-...
Parvez Hossain's user avatar
3 votes
1 answer
3k views

I am processing a sequence in chunks, where the last chunk may be shorter, and would like to show progress bar showing the number of items. The straightforward approach is import tqdm, math total=567 ...
eudoxos's user avatar
  • 19.2k
1 vote
0 answers
204 views

I am running react project developed using CRA in localhost in network tab it shows chunk.js files like shown in image I want it show original file names like which component file is rendering, so is ...
Akash Verma's user avatar
1 vote
2 answers
2k views

How to read a large file by chunking it and process each chunk sequentially then overwrite the resulted chunk to where it exactly came from(the same position or offset of file)? e.g: i want to read 1 ...
mohammad's user avatar
0 votes
1 answer
67 views

Consider a postgres table where for the date 2022-05-01 we have 200 values for various times: time value ...
sci9's user avatar
  • 776
1 vote
0 answers
145 views

I have a bunch of fit.objects which produce *.rds files larger than 100MB. GitHub comes with a file size limitation of 100MB and the solution of git lfs doesn't fit me since I don't want to pay for ...
mugdi's user avatar
  • 435
0 votes
1 answer
315 views

I have a large list with size of approx. 1.3GB. I'm looking for the fastest solution in R to generate chunks and save them in any convenient format so that : a) every saved file of the chunk is less ...
mugdi's user avatar
  • 435
-1 votes
1 answer
356 views

how can insert 1000000 row from textarea into database in laravel 8 ??????? i write this code and just can insert 30000 row and then browser give me HTTP ERROR 500 i set max_execution_time to 300 in ...
ProSonic's user avatar
2 votes
0 answers
2k views

I am trying to run a simple calculation based on two big gridded datasets in xarray (around 5 GB altogether, daily data from 1850-2100). I keep running out of memory when I try it this way import ...
scriptgirl_3000's user avatar
0 votes
0 answers
261 views

I have a large program where I am trying to read approximately 30000 lines of data and process it. I know that I can use the chuncksize functionality to do this, but I think I am not executing this ...
lcfields's user avatar
0 votes
1 answer
259 views

I have an API built in lumen and I plan to consume the API json response in the frontend using a single page framework like Angular. The problem is the response from some routes contain huge amount of ...
Nelson Thembeni's user avatar
3 votes
0 answers
559 views

I am reading this article about chunking a large database operation. I am also using django-import-export and django-import-export-celery in my admin site and I would like to integrate chunking into ...
Prosy A.'s user avatar
  • 2,822
0 votes
2 answers
3k views

Before marking this as answered by another question please note this is an array of arrays, not a flat array, also, the number I have given are an example, I have just shown them so you can visually ...
Web Nexus's user avatar
  • 1,148
0 votes
1 answer
527 views

I have a very large text file, and a function that does what I want it to do to each line. However, when reading line by line and applying the function, it takes roughly three hours. I'm wondering if ...
chrislee's user avatar
2 votes
0 answers
1k views

I have a block of code that looks like this, entities = self.session.query(Entities).filter(Entities.parent_id == 0) index_data = {} for entity in entities: data = entity.__dict__ data['...
Minura Punchihewa's user avatar
1 vote
1 answer
2k views

Basically my understanding is this: whenever a video player is playing media, it is downloading it in chunks, defined by the RANGE header. The server serves only the bytes requested from the file. ...
php_nub_qq's user avatar
  • 16.1k
0 votes
0 answers
357 views

I would like to upload files between 100-200mb using Flask on PythonAnywhere (which has a 100mb upload limit). I am trying to implement chunking but am still getting a 413 Request Entity Too Large ...
zweiss's user avatar
  • 1

1
2 3 4 5
8