Skip to content

bendbennett/go-api-demo

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

12 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Go API Demo

This repo contains code for an API written in Go.

Overview

The API is a Go implementation of Clean Architecture.

Stress testing is combined with the usage of metrics and tracing to compare the performance and behaviour when using an API backed with either an in-memory or relational data storage.

Change Data Capture (CDC) is used as a basis for event-driven cache and search engine population whilst avoiding the problem of dual writes.

Tags

Tag Implementation
v0.1.0 Basic HTTP and gRPC server exposing hello world endpoints.
v0.2.0 Adds HTTP and gRPC endpoints to create users stored in-memory.
v0.3.0 Stores created users either in-memory or in MySQL.
v0.4.0 Adds HTTP and gRPC endpoints to retrieve/read users.
v0.5.0
  • Adds Prometheus metrics and Grafana dashboard visualisation for HTTP requests.
  • Includes k6 script to stress-test HTTP requests to retrieve/read users and provides a comparison of retrieving users from in-memory storage and MySQL.
v0.6.0
  • Adds tracing with Jaeger to provide additional insight into slow request processing and errors observed when stress testing HTTP requests to retrieve/read users.
  • Adds metrics for gRPC requests.
v0.7.0 Adds Change Data Capture to publish events into Kafka when MySQL is mutated.
v0.8.0
  • Adds a Kafka consumer which populates Redis by processing Kafka events generated when the MySQL users table is mutated.
  • Includes stress-testing of HTTP requests to retrieve/read users, comparing performance of retrieving users from Redis and MySQL.
v0.9.0
  • Adds a Kafka consumer which populates Elasticsearch by processing Kafka events generated when the MySQL users table is mutated.
  • Adds HTTP and gRPC endpoints for searching users by name.
v0.10.0
  • Adds tracing for Redis, Elasticsearch and the Kafka consumers.
  • Switch to kafka-go.
  • Adds basic metrics and dashboard for the Kafka consumers.
v0.11.0
  • Uses 2 Kafka partitions for CDC for MySQL users table.
  • Uses 2 consumers for populating Elasticsearch.
v0.12.0 Uses Avro Schema for serialization and deserialization.

Set-up

To run the API you'll need to have Go and Docker installed.

gRPC

Install a Protocol Buffer Compiler, and the Go Plugins for the compiler (see the gRPC Quickstart for details) if you want to:

  • compile the .proto files by running make proto and/or
  • manually issue gRPC requests using gRPCurl.

Make

The Makefile contains commands for building, running and testing the API.

  • make run builds and runs the binary.
  • make build just builds the binary.
  • make fmt formats the code, updates import lines and runs clean.
  • make lint runs golangci-lint.
  • make proto generates protobuf files from proto files.
  • make test runs the linter then the tests (see Tests).
  • make migrate-up runs the database migrations for MySQL (see v0.3.0).

Tests

Install testify then run

make docker-up
make test

Manual Testing

You can test the API manually using a client. For instance Insomnia supports both HTTP and gRPC.

Alternatively, requests can be issued using cURL and gRPCurl (see v0.2.0, v0.3.0, v0.4.0).

v0.12.0

Uses Avro Schema for the serialization of users during CDC and deserialization when the Kafka consumers populate Redis and Elasticsearch.

Set-up

There is a delay between running docker-up and all necessary infrastructure being available to run the tests and the API. This delay can result in errors when running the integration tests. Execute make run once the tests run successfully.

make docker-up
make test
make run

The schema are visible through Kowl.

kowl_schema_registry

v0.11.0

Adds 2 Kafka partitions for CDC of MySQL users table.

Adds 2 Kafka consumers for populating Elasticsearch.

Set-up

There is a delay between running docker-up and all necessary infrastructure being available to run the tests and the API. This delay can result in errors when running the integration tests. Execute make run once the tests run successfully.

make docker-up
make test
make run

Running (see docker or local) the k6 script will send 50 requests per second (RPS) to the POST /user endpoint for 5 minutes.

Docker

 docker run -e HOST=host.docker.internal -i loadimpact/k6 run - <k6/post.js

Local

Install k6 and run:

k6 run -e HOST=localhost k6/post.js

Load Testing

user_post_consumer_2

The consumers populating Redis and Elasticsearch are able to process Kafka events at a sufficient rate to prevent any lag.

There are a couple of spikes in the 99th percentile request duration that warrant further investigation and/or profiling.

k6 Output

  execution: local
     script: k6/post.js
     output: -

  scenarios: (100.00%) 1 scenario, 100 max VUs, 5m30s max duration (incl. graceful stop):
           * constant_request_rate: 50.00 iterations/s for 5m0s (maxVUs: 20-100, gracefulStop: 30s)

running (5m00.0s), 000/046 VUs, 14974 complete and 0 interrupted iterations
constant_request_rate ✓ [======================================] 000/046 VUs  5m0s  50 iters/s

     ✓ status was 200

     checks.........................: 100.00% ✓ 14974     ✗ 0
     data_received..................: 3.5 MB  12 kB/s
     data_sent......................: 2.2 MB  7.3 kB/s
     dropped_iterations.............: 26      0.086671/s
     http_req_blocked...............: avg=7.55µs  min=3µs    med=5µs     max=1.19ms p(90)=8µs     p(95)=10µs
     http_req_connecting............: avg=965ns   min=0s     med=0s      max=1.11ms p(90)=0s      p(95)=0s
     http_req_duration..............: avg=24.04ms min=7.22ms med=14.54ms max=1.29s  p(90)=31.68ms p(95)=49.94ms
       { expected_response:true }...: avg=24.04ms min=7.22ms med=14.54ms max=1.29s  p(90)=31.68ms p(95)=49.94ms
     http_req_failed................: 0.00%   ✓ 0         ✗ 14974
     http_req_receiving.............: avg=80.84µs min=19µs   med=73µs    max=1.44ms p(90)=119µs   p(95)=138µs
     http_req_sending...............: avg=34.35µs min=13µs   med=31µs    max=693µs  p(90)=47µs    p(95)=59µs
     http_req_tls_handshaking.......: avg=0s      min=0s     med=0s      max=0s     p(90)=0s      p(95)=0s
     http_req_waiting...............: avg=23.93ms min=7.03ms med=14.43ms max=1.29s  p(90)=31.59ms p(95)=49.79ms
     http_reqs......................: 14974   49.915646/s
     iteration_duration.............: avg=24.29ms min=7.47ms med=14.78ms max=1.29s  p(90)=31.91ms p(95)=50.2ms
     iterations.....................: 14974   49.915646/s
     vus............................: 46      min=20      max=46
     vus_max........................: 46      min=20      max=46

v0.10.0

Adds tracing for Redis, Elasticsearch and the Kafka consumers that populate these data stores.

Adds basic metrics for the Kafka consumers.

Set-up

make docker-up
make run

Running (see docker or local) the k6 script will send 50 requests per second (RPS) to the POST /user endpoint for 5 minutes.

Docker

 docker run -e HOST=host.docker.internal -i loadimpact/k6 run - <k6/post.js

Local

Install k6 and run:

k6 run -e HOST=localhost k6/post.js

Load Testing

user_post_consumer_1

At the outset of the load test, the message consumption rate, consumer (msg/sec), for the consumer that populates Redis (Redis consumer - orange line) maintains throughput aligned with the rate at which users are being created. However, the message consumption rate for the consumer that populates Elasticsearch (Elasticsearch consumer - yellow line) has a noticeably lower throughput. The percentage of the queue capacity (100 messages) in use, suggests that there is a bottleneck in the processing of messages by the Elasticsearch consumer as it is continuously 100%. This is confirmed by the message lag which shows that Elasticsearch consumer processes messages less quickly than they are produced.

After 5 minutes, the requests being sent to the POST /user endpoint and the processing of messages by the Redis consumer cease. This results in an increase in the number of messages processed per second by the Elasticsearch consumer and, suggests that the bottleneck might have additional contributory factors.

v0.9.0

Adds a consumer which populates Elasticsearch by processing the Kafka events that are generated by Change Data Capture (CDC) when the MySQL users table is mutated.

Adds HTTP and gRPC endpoints for searching users by name.

Set-up

make docker-up
make run
make test

HTTP

Request

curl -i --request GET \
--url http://localhost:3000/user/search/ith
Response
HTTP/1.1 200 OK
Content-Type: application/json
Date: Mon, 08 Nov 2021 18:14:08 GMT
Content-Length: 261

[
  {
    "id":"39965d61-01f2-4d6d-8215-4bb88ef2a837",
    "first_name":"john",
    "last_name":"smith",
    "created_at":"2021-07-26T19:23:52Z"
  },
  {
    "id":"c0e137e3-6689-41c3-a421-0bf44f44d746",
    "first_name":"joanna",
    "last_name":"smithson",
    "created_at":"2021-07-26T19:23:52Z"
  }
]

gRPC

You'll need to generate a protoset and have gRPCurl installed.

Generate protoset

protoc \
-I=proto \
--descriptor_set_out=generated/user.protoset \
user.proto

Request

grpcurl \
-plaintext \
-protoset generated/user.protoset \
-d '{"searchTerm": "ith"}' \
localhost:1234 User/Search

Response

{
  "users": [
    {
      "id": "39965d61-01f2-4d6d-8215-4bb88ef2a837",
      "firstName": "john",
      "lastName": "smith",
      "createdAt": "2021-07-26T19:23:52Z"
    },
    {
      "id": "c0e137e3-6689-41c3-a421-0bf44f44d746",
      "firstName": "joanna",
      "lastName": "smithson",
      "createdAt": "2021-07-26T19:23:52Z"
    }
  ]
}

v0.8.0

Adds a consumer which populates Redis by processing the Kafka events that are generated by Change Data Capture (CDC) when the MySQL users table is mutated.

These changes are an implementation of event sourced CQRS, in that requests sent via HTTP or gRPC to create users mutate the users table in MySQL. These mutations are detected by CDC and result in events being published into Kafka. Consumer(s) process the Kafka events and create key-value entries in Redis. Requests to get users via HTTP or gRPC then retrieve users from Redis.

Set-up

make docker-up
make run

Load Testing

Following is the output from running the k6 script which ramps up from 1-200 virtual users (vus) over 2 minutes, maintains 200 vus for 1 minute, then ramps down to 0 vus over 2 minutes.

The two load tests were run using either Redis or MySQL data storage as described previously.

Redis

user_get_redis

16:40 - 16:45 marks the duration of the load test.

k6 Output
  execution: local
     script: k6/get.js
     output: -

  scenarios: (100.00%) 1 scenario, 200 max VUs, 5m30s max duration (incl. graceful stop):
           * default: Up to 200 looping VUs for 5m0s over 3 stages (gracefulRampDown: 30s, gracefulStop: 30s)


running (5m00.0s), 000/200 VUs, 530316 complete and 0 interrupted iterations
default ✓ [======================================] 000/200 VUs  5m0s

     ✓ status was 200

     checks.........................: 100.00% ✓ 530316      ✗ 0     
     data_received..................: 196 MB  654 kB/s
     data_sent......................: 45 MB   149 kB/s
     http_req_blocked...............: avg=3.91µs  min=1µs    med=3µs     max=2.54ms  p(90)=5µs      p(95)=6µs     
     http_req_connecting............: avg=133ns   min=0s     med=0s      max=1.33ms  p(90)=0s       p(95)=0s      
     http_req_duration..............: avg=67.85ms min=2.53ms med=69.18ms max=1.03s   p(90)=109.67ms p(95)=121.53ms
       { expected_response:true }...: avg=67.85ms min=2.53ms med=69.18ms max=1.03s   p(90)=109.67ms p(95)=121.53ms
     http_req_failed................: 0.00%   ✓ 0           ✗ 530316
     http_req_receiving.............: avg=51.82µs min=13µs   med=39µs    max=9.58ms  p(90)=80µs     p(95)=106µs   
     http_req_sending...............: avg=18.24µs min=5µs    med=14µs    max=20.29ms p(90)=26µs     p(95)=36µs    
     http_req_tls_handshaking.......: avg=0s      min=0s     med=0s      max=0s      p(90)=0s       p(95)=0s      
     http_req_waiting...............: avg=67.78ms min=2.49ms med=69.12ms max=1.03s   p(90)=109.59ms p(95)=121.46ms
     http_reqs......................: 530316  1767.762898/s
     iteration_duration.............: avg=67.99ms min=2.64ms med=69.32ms max=1.03s   p(90)=109.81ms p(95)=121.68ms
     iterations.....................: 530316  1767.762898/s
     vus............................: 1       min=1         max=200 
     vus_max........................: 200     min=200       max=200 

MySQL

user_get_mysql_2

17:10 - 17:15 marks the duration of the load test.

k6 Output
  execution: local
     script: k6/get.js
     output: -

  scenarios: (100.00%) 1 scenario, 200 max VUs, 5m30s max duration (incl. graceful stop):
           * default: Up to 200 looping VUs for 5m0s over 3 stages (gracefulRampDown: 30s, gracefulStop: 30s)


running (5m00.0s), 000/200 VUs, 230462 complete and 0 interrupted iterations
default ✓ [======================================] 000/200 VUs  5m0s

     ✗ status was 200
      ↳  95% — ✓ 220630 / ✗ 9832

     checks.........................: 95.73% ✓ 220630    ✗ 9832  
     data_received..................: 81 MB  270 kB/s
     data_sent......................: 19 MB  65 kB/s
     http_req_blocked...............: avg=84.71µs  min=1µs    med=3µs     max=1.07s   p(90)=5µs      p(95)=6µs     
     http_req_connecting............: avg=80.97µs  min=0s     med=0s      max=1.07s   p(90)=0s       p(95)=0s      
     http_req_duration..............: avg=156.47ms min=1.03ms med=27.59ms max=3s      p(90)=499.14ms p(95)=819.25ms
       { expected_response:true }...: avg=148.03ms min=1.03ms med=24.72ms max=2.83s   p(90)=473.49ms p(95)=789.24ms
     http_req_failed................: 4.26%  ✓ 9832      ✗ 220630
     http_req_receiving.............: avg=52.43µs  min=13µs   med=43µs    max=12.94ms p(90)=81µs     p(95)=103µs   
     http_req_sending...............: avg=18.74µs  min=5µs    med=15µs    max=21.92ms p(90)=26µs     p(95)=34µs    
     http_req_tls_handshaking.......: avg=0s       min=0s     med=0s      max=0s      p(90)=0s       p(95)=0s      
     http_req_waiting...............: avg=156.4ms  min=953µs  med=27.51ms max=3s      p(90)=499.07ms p(95)=819.22ms
     http_reqs......................: 230462 768.22389/s
     iteration_duration.............: avg=156.69ms min=1.12ms med=27.74ms max=3s      p(90)=499.67ms p(95)=819.84ms
     iterations.....................: 230462 768.22389/s
     vus............................: 1      min=1       max=200 
     vus_max........................: 200    min=200     max=200 

v0.7.0

Adds Change Data Capture (CDC) using the Debezium MySQL Connector to publish messages into Kafka when the underlying MySQL database is mutated.

Set-up

make docker-up

Kowl (GUI for Kafka) is accessible at http://localhost:8080/.

Kafka Topics & Messages

Initially, 5 topics will be visible if the database migrations have not been previously run.

kowl_initial_topics

Additional topics will be visible following mutations of the database, for example, creation of users, either through:

  • make test or,
  • using the cURL and/or gRPCurl requests for create user endpoints (see v0.2.0).

kowl_users_migrations_topics

kowl_users_messages

v0.6.0

Adds tracing using Jaeger (see Quick Start) and further metrics around gRPC requests.

Set-up

make docker-up

Jaeger is accessible at http://localhost:16686.

Tracing

Overview

Using the cURL and/or gRPCurl requests for user endpoints (see v0.2.0, v0.3.0 and v0.4.0) will generate traces.

The following illustrates traces generated from:

  • gRPC request to /User/Read endpoint
  • HTTP request to GET /user endpoint
  • MySQL ping during application boot

jaeger_user_get_http_grpc

Individual traces

Drilling into the HTTP request reveals further information about where time is spent. This particular request takes 5.16 ms, 4.15 ms (~80%) of which is consumed by the SQL query.

jaeger_user_get_http_detail

Logging and Tracing with Load Testing

Slow Requests

Running the k6 script (see docker or local) reveals that as the number of requests increases the length of time spent trying to obtain a SQL connection increases.

jaeger_user_get_http_slow

Failures

As the number of requests increases further failures are visible both in the logs and in the corresponding traces.

Error - Too many connections
{
  "level": "error",
  "ts": 1628240198.250581,
  "caller": "log/spanlogger.go:35",
  "msg": "Error 1040: Too many connections",
  "commit_hash": "a94f733b84fe0f102e658f8a6a77a512d57476fb",
  "trace_id": "466ea0fc21ccdf27",
  "span_id": "466ea0fc21ccdf27",
  "stacktrace": "github.com/bendbennett/go-api-demo/internal/log.spanLogger.Error...."
}

jaeger_user_get_http_connection_error

Error - Context deadline exceeded
{
  "level": "error",
  "ts": 1628240198.376878,
  "caller": "log/spanlogger.go:35",
  "msg": "context deadline exceeded",
  "commit_hash": "a94f733b84fe0f102e658f8a6a77a512d57476fb",
  "trace_id": "5875ddde4e48b9c2",
  "span_id": "5875ddde4e48b9c2",
  "stacktrace": "github.com/bendbennett/go-api-demo/internal/log.spanLogger.Error...."
}

jaeger_user_get_deadline_exceeded_error

v0.5.0

Adds metrics and dashboard visualisation for HTTP request duration, using Prometheus and Grafana, respectively.

Set-up

make docker-up

Prometheus is accessible at http://localhost:9090.

Grafana is accessible at http://localhost:3456

  • The dashboard for requests can be found by using Search and drilling into the Requests folder.

Metrics

Using the cURL requests for user endpoints (see v0.2.0 and v0.3.0) will generate metrics.

Raw metrics can be seen by running curl localhost:3000/metrics.

Metrics are also visible in Prometheus and HTTP request duration can be seen by searching for:

  • request_duration_seconds_bucket
  • request_duration_seconds_count
  • request_duration_seconds_sum

Load Testing

Running (see docker or local) the k6 script will generate meaningful output for the Grafana dashboard.

Docker

 docker run -e HOST=host.docker.internal -i loadimpact/k6 run - <k6/get.js

Local

Install k6 and run:

k6 run -e HOST=localhost k6/get.js

Dashboard

Following is the output from running the k6 script which ramps up from 1-200 virtual users (vus) over 2 minutes, maintains 200 vus for 1 minute, then ramps down to 0 vus over 2 minutes.

The two load tests were run using either MySQL or in-memory data storage (see v0.3.0 for configuration).

MySQL

user_get_mysql

08:55 - 09:00 marks the duration of the load test.

Initially the number of requests per second (rps) increases as the number of vus rises,
and, an accompanying increase in the request duration can be seen in the heat map for successful (200) requests.

The rps then decreases and more of the successful requests take longer, plateauing during the sustained load of 200 vus and is accompanied by the emergence of failed (500) requests.

Logs
{"commitHash":"e04cc2d0917ead700130dd378376a75a21c99930","level":"warning","msg":"context deadline exceeded","time":"2021-07-30T08:58:03+01:00"}
{"commitHash":"e04cc2d0917ead700130dd378376a75a21c99930","level":"warning","msg":"Error 1040: Too many connections","time":"2021-07-30T08:58:04+01:00"}

As the number of vus is ramped down, rps increases, successful request duration decreases and failed requests disappear.

k6 Output
  execution: local
     script: k6/get.js
     output: -

  scenarios: (100.00%) 1 scenario, 200 max VUs, 5m30s max duration (incl. graceful stop):
           * default: Up to 200 looping VUs for 5m0s over 3 stages (gracefulRampDown: 30s, gracefulStop: 30s)


running (5m00.0s), 000/200 VUs, 141654 complete and 0 interrupted iterations
default ✓ [======================================] 000/200 VUs  5m0s

     ✗ status was 200
      ↳  96% — ✓ 136806 / ✗ 4848

     checks.........................: 96.57% ✓ 136806     ✗ 4848
     data_received..................: 50 MB  167 kB/s
     data_sent......................: 12 MB  40 kB/s
     http_req_blocked...............: avg=27.32µs  min=1µs    med=4µs     max=325.8ms  p(90)=6µs      p(95)=7µs
     http_req_connecting............: avg=22.56µs  min=0s     med=0s      max=325.72ms p(90)=0s       p(95)=0s
     http_req_duration..............: avg=254.72ms min=1.81ms med=76.84ms max=3.02s    p(90)=794.9ms  p(95)=1.1s
       { expected_response:true }...: avg=246.55ms min=1.81ms med=71.32ms max=2.78s    p(90)=795.03ms p(95)=1.11s
     http_req_failed................: 3.42%  ✓ 4848       ✗ 136806
     http_req_receiving.............: avg=66.19µs  min=17µs   med=57µs    max=29.42ms  p(90)=96µs     p(95)=122µs
     http_req_sending...............: avg=23.78µs  min=6µs    med=20µs    max=31.37ms  p(90)=33µs     p(95)=41µs
     http_req_tls_handshaking.......: avg=0s       min=0s     med=0s      max=0s       p(90)=0s       p(95)=0s
     http_req_waiting...............: avg=254.63ms min=1.75ms med=76.75ms max=3.02s    p(90)=794.82ms p(95)=1.1s
     http_reqs......................: 141654 472.187596/s
     iteration_duration.............: avg=254.92ms min=1.91ms med=77.05ms max=3.02s    p(90)=795.13ms p(95)=1.1s
     iterations.....................: 141654 472.187596/s
     vus............................: 1      min=1        max=200
     vus_max........................: 200    min=200      max=200

In-Memory

user_get_in_memory

08:00 - 08:05 marks the duration of the load test.

Initially the number of requests per second (rps) increases as the number of vus rises, and, successful (200) request duration remains stable.

The rps levels off with successful request duration remaining stable.

As the number of vus is ramped down, rps remains stable. There are no failed (500) requests during this load test.

k6 Output
  execution: local
     script: k6/get.js
     output: -

  scenarios: (100.00%) 1 scenario, 200 max VUs, 5m30s max duration (incl. graceful stop):
           * default: Up to 200 looping VUs for 5m0s over 3 stages (gracefulRampDown: 30s, gracefulStop: 30s)


running (5m00.0s), 000/200 VUs, 1816350 complete and 0 interrupted iterations
default ✓ [======================================] 000/200 VUs  5m0s

     ✓ status was 200

     checks.........................: 100.00% ✓ 1816350     ✗ 0
     data_received..................: 200 MB  666 kB/s
     data_sent......................: 153 MB  509 kB/s
     http_req_blocked...............: avg=3.66µs  min=1µs      med=3µs    max=20.91ms p(90)=4µs     p(95)=5µs
     http_req_connecting............: avg=41ns    min=0s       med=0s     max=7.8ms   p(90)=0s      p(95)=0s
     http_req_duration..............: avg=19.72ms min=235µs    med=2.73ms max=2.34s   p(90)=13.12ms p(95)=32.86ms
       { expected_response:true }...: avg=19.72ms min=235µs    med=2.73ms max=2.34s   p(90)=13.12ms p(95)=32.86ms
     http_req_failed................: 0.00%   ✓ 0           ✗ 1816350
     http_req_receiving.............: avg=43.5µs  min=12µs     med=35µs   max=17.14ms p(90)=58µs    p(95)=76µs
     http_req_sending...............: avg=17.17µs min=5µs      med=13µs   max=19.74ms p(90)=22µs    p(95)=29µs
     http_req_tls_handshaking.......: avg=0s      min=0s       med=0s     max=0s      p(90)=0s      p(95)=0s
     http_req_waiting...............: avg=19.65ms min=208µs    med=2.67ms max=2.34s   p(90)=13.05ms p(95)=32.78ms
     http_reqs......................: 1816350 6054.831623/s
     iteration_duration.............: avg=19.85ms min=289.96µs med=2.86ms max=2.34s   p(90)=13.26ms p(95)=33.11ms
     iterations.....................: 1816350 6054.831623/s
     vus............................: 1       min=1         max=200
     vus_max........................: 200     min=200       max=200

v0.4.0

Adding HTTP and gRPC endpoints for retrieving users.

HTTP

Request

curl -i --request GET \
--url http://localhost:3000/user
Response
HTTP/1.1 200 OK
Content-Type: application/json
Date: Mon, 26 Jul 2021 19:29:12 GMT
Content-Length: 251

[
  {
    "id":"39965d61-01f2-4d6d-8215-4bb88ef2a837",
    "first_name":"john",
    "last_name":"smith",
    "created_at":"2021-07-26T19:23:52Z"
  },
  {
    "id":"c0e137e3-6689-41c3-a421-0bf44f44d746",
    "first_name":"joanna",
    "last_name":"smithson",
    "created_at":"2021-07-26T19:23:52Z"
  }
]

gRPC

You'll need to generate a protoset and have gRPCurl installed.

Generate protoset

protoc \
-I=proto \
--descriptor_set_out=generated/user.protoset \
user.proto

Request

grpcurl \
-plaintext \
-protoset generated/user.protoset \
localhost:1234 User/Read

Response

{
  "users": [
    {
      "id": "39965d61-01f2-4d6d-8215-4bb88ef2a837",
      "firstName": "john",
      "lastName": "smith",
      "createdAt": "2021-07-26T19:23:52Z"
    },
    {
      "id": "c0e137e3-6689-41c3-a421-0bf44f44d746",
      "firstName": "joanna",
      "lastName": "smithson",
      "createdAt": "2021-07-26T19:23:52Z"
    }
  ]
}

v0.3.0

Stores created users either in-memory or in MySQL.

To use MySQL storage you'll need to install golang-migrate and run the following commands:

make docker-up
make migrate-up

Running make docker-up will

  • copy .env.dist => .env
    • USER_STORAGE (either memory or sql) determines whether users are stored in-memory or in MySQL.
  • start a docker-based instance of MySQL.

Running make migrate-up creates the table in MySQL for storing users.

The same cURL and gRPCurl requests as described for v0.2.0 can be used.

v0.2.0

Adding HTTP and gRPC endpoints for user creation.

Users are stored in-memory.

HTTP

Request

curl -i --request POST \
--url http://localhost:3000/user \
--header 'Content-Type: application/json' \
--data '{
    "first_name": "john",
    "last_name": "smith"
}'
Response
HTTP/1.1 201 Created
Content-Type: application/json
Date: Tue, 06 Jul 2021 12:03:25 GMT
Content-Length: 127

{
    "id":"afaa2920-77e4-49d0-a29f-5f2d9b6bf2d1",
    "first_name":"john",
    "last_name":"smith",
    "created_at":"2021-07-06T13:03:25+01:00"
}

gRPC

You'll need to generate a protoset and have gRPCurl installed.

Generate protoset

protoc \
-I=proto \
--descriptor_set_out=generated/user.protoset \
user.proto

Request

grpcurl \
-plaintext \
-protoset generated/user.protoset \
-d '{"first_name": "john", "last_name": "smith"}' \
localhost:1234 User/Create

Response

{
    "id": "ca3d9549-eb8d-4b5a-b45f-3551fb4fbdc9",
    "firstName": "john",
    "lastName": "smith",
    "createdAt": "2021-07-06T13:08:45+01:00"
}

v0.1.0

Basic HTTP and gRPC server.

HTTP

Request

curl -i localhost:3000
Response
HTTP/1.1 200 OK
Date: Tue, 22 Jun 2021 11:19:48 GMT
Content-Length: 0

gRPC

You'll need to generate a protoset and have gRPCurl installed.

Generate protoset

protoc \
-I=proto \
--descriptor_set_out=generated/hello.protoset \
hello.proto

Request

grpcurl \
-plaintext \
-protoset generated/hello.protoset \
-d '{"name": "world"}' \
localhost:1234 Hello/Hello

Response

{
  "message": "Hello world"
}

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Packages

No packages published