stream-processing
Here are 813 public repositories matching this topic...
-
Updated
Jun 21, 2022 - Java
-
Updated
Jun 24, 2022
-
Updated
Jun 7, 2022 - Python
-
Updated
Jun 29, 2022
Is your feature request related to a problem? Please describe.
A clear and concise description of what the problem is. Ex. I'm always frustrated when [...]
Currently, ksqlDB causes a full topic scan whenever performing a pull query over a stream. This is inefficient when looking up specific sets of data, but necessary due to how pull queries are implemented over streams.
**Describe the
-
Updated
Jul 8, 2022 - Go
For an implementation of #126 (PostgreSQL driver with SKIP LOCKED), I create a SQL table for each consumer group containing the offsets ready to be consumed. The name for these tables is build by concatenating some prefix, the name of the topic and the name of the consumer group. In some of the test cases in the test suite, UUID are used for both, the topic and the consumer group. Each UUID has
-
Updated
Jul 10, 2022 - Rust
-
Updated
Jul 9, 2022 - C
to_dict() equivalent
I would like to convert a DataFrame to a JSON object the same way that Pandas does with to_dict().
toJSON() treats rows as elements in an array, and ignores the index labels. But to_dict() uses the index as keys.
Here is an example of what I have in mind:
function to_dict(df) {
const rows = df.toJSON();
const entries = df.index.map((e, i) => ({ [e]: rows[i] }));
-
Updated
Jul 8, 2022 - Python
-
Updated
Jul 10, 2022 - Java
now our hash join can only handle the equal condition of two input columns, but can not handle the equal conditions of expression.
dev=> explain select * from t1,t2 where t1.v1 = t2.v1 + t2.v2;
QUERY PLAN
---------------------------------------------------------------------------------------
BatchNestedLoopJoin
-
Updated
Feb 20, 2022 - C
-
Updated
Jan 4, 2022 - Go
-
Updated
Jul 9, 2022
-
Updated
Dec 28, 2021 - Python
It can be very difficult to piece together a reasonably estimate of a history of events from the current workers logs because none of them have timestamps.
So for that end, I think we should add timestamps to the logs.
This has some cons:
- We can't just use
@printflike we have been until now. We need to either include a timestamp in every@printfcall (laborious and error prone) or c
-
Updated
May 20, 2022 - Java
-
Updated
Jul 4, 2022 - Go
For example, given a simple pipeline such as:
Pipeline p = Pipeline.create();
p.readFrom(TestSources.items("the", "quick", "brown", "fox"))
.aggregate(aggregator)
.writeTo(Sinks.logger());
I'd like aggregator to be something requiring a non-serialisable dependency to do its work.
I know I can do this:
Pipeline p = Pipeline.create();
p.readFrom(TestSource
-
Updated
Jun 1, 2022 - Go
-
Updated
Jul 1, 2022 - JavaScript
The mapcat function seems to choke if you pass in a mapping function that returns a stream instead of a sequence:
user> (s/stream->seq (s/mapcat (fn [x] (s/->source [x])) (s/->source [1 2 3])))
()
Aug 18, 2019 2:23:39 PM clojure.tools.logging$eval5577$fn__5581 invoke
SEVERE: error in message propagation
java.lang.IllegalArgumentException: Don't know how to create ISeq from: manifold.
-
Updated
Jul 8, 2022 - Java
-
Updated
Jul 1, 2022 - Java
Improve this page
Add a description, image, and links to the stream-processing topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the stream-processing topic, visit your repo's landing page and select "manage topics."

I figured out a way to get the (x,y,z) data points for each frame from one hand previously. but im not sure how to do that for the new holistic model that they released. I am trying to get the all landmark data points for both hands as well as parts of the chest and face. does anyone know how to extract the holistic landmark data/print it to a text file? or at least give me some directions as to h