You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/backend/spark-performance.md
+6-6Lines changed: 6 additions & 6 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -14,7 +14,7 @@ All tests run here were using [Bullet-Spark 0.1.2](https://github.com/bullet-db/
14
14
15
15
*[jq](https://stedolan.github.io/jq/) - a nice tool to parse Bullet JSON responses
16
16
* curl, bash and python - for running and analyzing Bullet queries
17
-
*[Apache JMeter](https://jmeter.apache.org/) - a load testing tool to send queries to the server simultaneously
17
+
*[Apache JMeter](https://jmeter.apache.org/) - a load testing tool to send multiple queries to the server simultaneously
18
18
19
19
### Cluster
20
20
@@ -112,11 +112,11 @@ Command line:
112
112
113
113
## Test 1: Latency of Bullet Spark
114
114
115
-
This test was done on the smaller data. We used a [RAW query without any filtering](../ws/examples/#simplest-query) to measure the latency added by Bullet Spark. This is not the end-to-end latency for a query. It is the latency from receiving the query to finishing the query, not includes the time spent in Kafka. We ran this query 100 times.
115
+
This test was done on the smaller data. We used a [RAW query without any filtering](../ws/examples/#simplest-query) to measure the latency added by Bullet Spark. This is not the end-to-end latency for a query. It is the latency from receiving the query to finishing the query, not including the time spent in Kafka. We ran this query 100 times.
@@ -126,7 +126,7 @@ The average latency was 1173 ms. This result shows that this is the fastest Bull
126
126
127
127
## Test 2: Scalability for smaller data
128
128
129
-
This test was done on the smaller data. We want to measure how many of queries we can have running simultaneously on Bullet Spark. We ran 400, 800, 1500 and 1100 queries each for 10 minutes.
129
+
This test was done on the smaller data. We want to measure how many queries we can have running simultaneously on Bullet Spark. We ran 400, 800, 1500 and 1100 queries each for 10 minutes.
130
130
131
131
### Result
132
132
@@ -168,7 +168,7 @@ This test was done on the smaller data. We want to measure how many of queries w
168
168
169
169
### Conclusion
170
170
171
-
The average processing time for each batch was 1 second 143 ms which was below the batch duration 2 seconds. For average, 1 CPU core and 3GB memory were used in this experiment. CPU and memory usages go slowly up while queries number goes up but they are still within resource limits. We can easily run up to 1500 RAW queries simultaneously in this test.
171
+
The average processing time for each batch was 1 second 143 ms which was below the batch duration 2 seconds. On average, 1 CPU core and 3GB memory were used in this experiment. CPU and memory usages go slowly up while queries number goes up but they are still within resource limits. We can easily run up to 1500 RAW queries simultaneously in this test.
172
172
173
173
## Test 3: Scalability for larger data
174
174
@@ -202,4 +202,4 @@ This test was done on the larger data. We ran 100, 400, 800 and 600 queries each
202
202
203
203
### Conclusion
204
204
205
-
The average processing time for each batch was 3 second 97 ms which was below the batch duration 5 seconds. For average, 1.2 CPU core and average 5GB memory were used in this experiment. But with queries number goes up, some of the executors memory usage were up to 8-10GB which is close to our resource limits. With more queries running, OOM may happen. So in this experiment, we can only afford up to 800 queries simultaneously.
205
+
The average processing time for each batch was 3 seconds 97 ms which was below the batch duration 5 seconds. On average, 1.2 CPU core and average 5GB memory were used in this experiment. But with queries number goes up, some of the executors memory usage were up to 8-10GB which is close to our resource limits. With more queries running, OOM may happen. So in this experiment, we can only afford up to 800 queries simultaneously.
0 commit comments