Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Unusual query timing results with aggregator connection with caching enabled when some connections have errors #4288

Open
cbruno10 opened this issue May 28, 2024 · 2 comments
Labels
question Further information is requested

Comments

@cbruno10
Copy link
Contributor

cbruno10 commented May 28, 2024

Describe the bug
When I query an aggregator with caching enabled, and I encounter errors, e.g.,

aws_pipeling_012: table 'aws_s3_bucket' column 'region' requires hydrate data from getBucketLocation, which failed with error operation error S3: GetBucketLocation, exceeded maximum number of attempts, 9, https response error StatusCode: 0, RequestID: , HostID: , request send failed, Get "https://stackset-test-bucket.s3.us-east-1.amazonaws.com/?location=": lookup stackset-test-bucket.s3.us-east-1.amazonaws.com on 192.168.1.1:53: read udp 192.168.1.22:62230->192.168.1.1:53: i/o timeout.

for some of the connections, the output shows the timing decreasing on subsequent queries but it doesn't report cached rows until the 2nd or further queries.

For instance, when running select name, region from aws_s3_bucket for an aggregator connection with 2 connections:

  • Time: 235.8s. Rows returned: 38. Rows fetched: 42. Hydrate calls: 42. Scans: 2. Connections: 2.
  • Time: 9.9s. Rows returned: 42. Rows fetched: 42. Hydrate calls: 42. Scans: 2. Connections: 2.
  • Time: 65ms. Rows returned: 42. Rows fetched: 42 (cached). Hydrate calls: 0. Scans: 2. Connections: 2.

With 15 connections:

  • Time: 489.9s. Rows returned: 307. Rows fetched: 722. Hydrate calls: 722. Scans: 15. Connections: 15.
  • Time: 534.0s. Rows returned: 389. Rows fetched: 722. Hydrate calls: 722. Scans: 15. Connections: 15.
  • Time: 279.7s. Rows returned: 687. Rows fetched: 722 (48 cached). Hydrate calls: 674. Scans: 15. Connections: 15.
  • Time: 394ms. Rows returned: 722. Rows fetched: 722 (482 cached). Hydrate calls: 240. Scans: 15. Connections: 15.
  • Time: 83ms. Rows returned: 722. Rows fetched: 722 (cached). Hydrate calls: 0. Scans: 15. Connections: 15.
  • Time: 87ms. Rows returned: 722. Rows fetched: 722 (cached). Hydrate calls: 0. Scans: 15. Connections: 15.
  • Time: 84ms. Rows returned: 722. Rows fetched: 722 (cached). Hydrate calls: 0. Scans: 15. Connections: 15.

In both cases, the timing starts to decrease on the 2nd run, and then even more on subsequent runs until all rows are cached.

In the 2nd run, are any cached rows/results being used (or maybe just some of the hydrate calls)? As the query time decreases, I was assuming that was from it using some cached info, but it was difficult to tell from the timing/rows info.

Steampipe version (steampipe -v)
Steampipe v0.23.2

To reproduce
Run the query above with at least 2 connections (and 20-30 buckets in each account)

Expected behavior
I'm not sure, maybe improvements to the rows fetched/hydrate calls info?

Additional context
Add any other context about the problem here.

@cbruno10 cbruno10 added bug Something isn't working question Further information is requested and removed bug Something isn't working labels May 28, 2024
@cbruno10 cbruno10 changed the title Unusual query timing results with aggregator connection with caching enabled when some queries have errors Unusual query timing results with aggregator connection with caching enabled when some connections have errors May 28, 2024
Copy link

This issue is stale because it has been open 60 days with no activity. Remove stale label or comment or this will be closed in 30 days.

@github-actions github-actions bot added the stale No recent activity has been detected on this issue/PR and it will be closed label Jul 28, 2024
Copy link

This issue was closed because it has been stalled for 90 days with no activity.

@github-actions github-actions bot closed this as not planned Won't fix, can't repro, duplicate, stale Aug 27, 2024
@pskrbasu pskrbasu reopened this Oct 1, 2024
@pskrbasu pskrbasu removed the stale No recent activity has been detected on this issue/PR and it will be closed label Oct 1, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

2 participants