In my GitHub Repository IBM FHIR Server, we use GitHub Actions to execute our Continuous Integration (CI) workflows. The CI workflows enable us to execute our code in complicated scenarios – a database (IBM Db2, Postgres), events system (Kafka (two zookeeper, two brokers), NATS) in a mix of various configurations.
I started to enable Docker Compose with Kafka and the IBM FHIR Server’s Audit module.
I ran into this error when running docker-compose exec in my workflow’s run.
TEST_CONFIGURATION: check that there is output and the configuration works
the input device is not a TTY
Error: Process completed with exit code 1.
Thanks for sitting down and watching this video. I’m going to show you how to quickly spin up a Docker image of IBM FHIR Server, check the logs, make sure it’s healthy, and how to use the fhir-examples module with the near search.
The following are the directions followed in the video:
Run the Server docker run -p9443:9443 ibmcom/ibm-fhir-server
Note, startup may take 2 minutes as the image is bootstrapping a new Apache Derby database in the image. To use Postgres or IBM Db2, please review the documentation.
Review the docker logs
Check the server is up and operational
curl -k -i -u 'fhiruser:change-password' 'https://localhost:9443/fhir-server/api/v4/$healthcheck'
You now have a running IBM FHIR Sever.
Let’s load some data using a Jupyter Notebook.
The IBM FHIR Server team wraps specification and service unit tests into a module called fhir-examples and posts to Bintray: ibm-fhir-server-releases or go directly to the repository.
We’re going to use the python features and Jupyter Notebook to process the fhir-examples.
We’ll download the zip, filter the interesting jsons, and upload to the IBM FHIR Server in a loop.
entries = z.namelist()
for entry in entries:
if entry.startswith('json/ibm/bulk-data/location/'):
f = z.open(entry);
content = f.read()
r = requests.post('https://localhost:9443/fhir-server/api/v4/Location',
data=content,
headers=headers,
auth=httpAuth,
verify=False)
print('Done uploading - ' + entry)
We’re going to query the data on the IBM FHIR Server using the Search Query Parameter near to search within 10Km of Cambridge Massachusetts.
The following takes the Capability Statements and generates the minimum coverage set of Resources across implementation guides based on the capability statements.
Resources and Profiles output in Markdown
This is handy to check the number of profiles in use (beyond the base spec) and to know the profiles used on the server based on the capability statements.
To check the operations rrequired in the implementation guides, you can use the following to process a set of capability statements into a useful outputs.
My team has been running more workloads on IBM Cloud, more specifically with IBM Db2. Our daily tools are slightly different to work with in the cloud, less administrative access and tools we can access on the host – db2batch, db2advis, db2expln and other native tools.
That’s when I ran across some great references that lead me in a direction that works for my team.
db2 update dbm cfg using SSL_CLNT_KEYDB \
/database/config/db2inst1/SSL_CLIENT/ibmca.kdb
db2 update dbm cfg using SSL_CLNT_STASH
/database/config/db2inst1/SSL_CLIENT/ibmca.sth
db2 update dbm cfg using keystore_location
/database/config/db2inst1/SSL_CLIENT/ibmca.kdb
Restart the database
db2stop
db2start
Catalog the database
db2 catalog tcpip node cdtdb1 remote \
dashdb-txn-flex-yp-xxxx-xxxx.services.dal.bluemix.net server 50001 security ssl
db2 catalog db bludb as fhirblu4 at node cdtdb1
db2 connect to fhirblu4 user testpaul using ^PASSWORD^
If you have a problem connecting, log out of db2inst1 and log back in. It’ll activate the db2profile again.
-amd keeps the build focused only on the necessary packages (not the full fhir-parent)
Idempotent Execution of the Role Creation
su - db2inst1 -c "db2 \"connect to fhirdb\" && db2 \" BEGIN IF (SELECT ROLENAME FROM SYSCAT.ROLES WHERE ROLENAME = 'FHIRSERVER') IS NULL THEN EXECUTE IMMEDIATE 'CREATE ROLE FHIRSERVER'; END IF; END;\""
su - db2inst1 -c "db2 \"connect to fhirdb\" && db2 \" BEGIN IF (SELECT ROLENAME FROM SYSCAT.ROLES WHERE ROLENAME = 'FHIRBATCH') IS NULL THEN EXECUTE IMMEDIATE 'CREATE ROLE FHIRBATCH'; END IF; END;\""
Shell Pipestatus
Checking the Status of any command in a pipe, it was helpful in some automation where I had to wait on a jar to finish and check the output. Source
docker run -p 8080:8080 --rm apache/nifi:latest bash
Find the docker container id
$ docker ps
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
09d2a7395fa2 apache/nifi:latest "../scripts/start.sh…" 7 seconds ago Up 6 seconds 8000/tcp, 8443/tcp, 10000/tcp, 0.0.0.0:8080->8080/tcp gracious_rosalind
Copy the fhirKeystore.p12 (in this case we just updated this one only).
Extracting a resource from a FHIR Bundle with over 10000 entries, and you know there is a problem at a specific resource, then you can use jq and array processing to extract the resource:
Thomas Alva Edison was a famous American inventor and businessman, “described as America’s greatest inventor”, and was one of the most prolific inventors in US history. Thomas Edison was granted/filed 1084 patents from 1847-1931.[1] He’s just one cool inventor – lamps, light bulbs, phonograph and so many more life changing inventions.
Google Patents has a wonderful depth of patent history, and the history is searchable with custom search strings:
inventor:(Thomas Edison) before:priority:19310101
inventor:(Paul R Bastide) after:priority:2009-01-01
Google provides a seriously cool feature – a downloadable csv. Pandas anyone? The content is provided in an agreement between the USPTO and Google. Google also provides it as part of the Google APIs/Platform. The data is fundamentally public, and Google has made it very accessible with some GitHub examples. [2] The older patent data more difficult to search as the content has been scraped from Optical Character Recognition.
I have found a cross-section of three things I am very interested in: History, Inventing and Data Science. Time to see what cool things about the Edison data.
Step
To start the playing with the data, one must install Jupyter.
Launch the Edison notebook and follow along with the cells.
The notebook renders some interesting insights using numpy, pandas, matplotlib and scipy. The notebook includes a cell to install python libraries, and once one executes the per-requisites cell; all is loaded.
The Jupyter notebook loads the data using an input cell, once run, the analytics enable me to see the number of co-inventors (but need to cleanse the data first).
One notices that Thomas Alva is not an inventor in those results, as such one needs to modify to the notebook to use the API with more recent Inventors. With the comprehensive APIs from USPTO, one extracts patent data by one of a number of JSON REST APIs. Kudos to the USPTO to really open up the data and the API.
Conclusion
All-in the APIs/Python/Jupyter Notebook/Analysis are for fun, and provide insight into Thomas Edison’s patent data – one focused individual.
References
[1] Prolific Inventors https://en.wikipedia.org/wiki/List_of_prolific_inventors number wise it appears to conflict with https://en.wikipedia.org/wiki/List_of_Edison_patents which reports 1093 (it’s inclusive of design patents) [2] Google / USPTO Patent Data https://www.google.com/googlebooks/uspto-patents.html [3] USPTO Open Data https://developer.uspto.gov/about-open-data and https://developer.uspto.gov/api-catalog
[4] PatentsView http://www.patentsview.org/api/faqs.html
To simply testing with Zookeeper on a remote Kafka cluster, one must connect to the client application ports on the backend. When the remote Kafka cluster has multiple nodes and behind a firewall and a SSH jump server, the complexity is fairly high. Note, the SSH jump server is the permitted man in the middle. The client must allow application access to Zookeeper on Kafka – listening locally. Current techniques allow for a single port hosted on the developers machine for instance, 2181 listening on the local machine, and a single remote server. This approach is not reliable – servers are taken out of service, added back, fail, or reroute to the master (another separate server).
3 – Save the hosts file 4 – Setup Available interfaces (1 for each unique service) 1 is already up and in use (you only need to add the extras)
sudo ifconfig lo0 alias 127.0.0.2 up
sudo ifconfig lo0 alias 127.0.0.3 up
sudo ifconfig lo0 alias 127.0.0.4 up
sudo ifconfig lo0 alias 127.0.0.5 up
5 – Setup the port forwarding, forward to jump server ssh -L 30991:localhost:30991 jump-server 6 – Forward to Kafka server ssh -L 30991:localhost:2181 kafka-1 7 – Loop while on kafka server while true; do echo “waiting”; sleep 180; done 8 – Repeat for each kafka server increasing the port by 1 (refer to ports section for mapping) 9 – Setup the Terminal – node krb5-tcp.js 10 – Setup the Terminal – node proxy_socket.js
echo stats | nc kafka-1 2181 Zookeeper version: 3.4.6-IBM_4–1, built on 06/17/2016 01:58 GMT Clients: /192.168.12.47:50404[1](queued=0,recved=1340009,sent=1360508) /192.168.12.46:48694[1](queued=0,recved=1348346,sent=1368936) /192.168.12.48:39842[1](queued=0,recved=1341655,sent=1362178) /0:0:0:0:0:0:0:1:39644[0](queued=0,recved=1,sent=0)
public class MyCallBackHandler implements CallbackHandler { public void handle(Callback[] callbacks) throws IOException, UnsupportedCallbackException {
for (int i = 0; i < callbacks.length; i++) { System.out.println(callbacks[i]); } } }
For a project I am working on I needed to rewrite a DGram port. I moved the ports around and found a few quick tests.
Testing with NC
my-machine:~$ echo -n “data-message” | nc -v -4u -w1 localhost 88
found 0 associations
found 1 connections:
1: flags=82<CONNECTED,PREFERRED>
outif (null)
src 127.0.0.1 port 53862
dst 127.0.0.1 port 88
rank info not available
Connection to localhost port 88 [udp/radan-http] succeeded!
Rewriting incoming datagrams to another port
You can run the sample, and get the results as follows