dependency-check is a standalone maven plugin which checks for vulnerable dependencies. It’s hosted on GitHub. I switched to it from the victims-db, which no longer looks like it is updated. I had to carefully analyze the output, it was very helpful finding one issue where we had an unintended include.
I recently attended the HL7 FHIR Connectathon 29. For those that are not familiar with Connectathons, I think they are fairly unique events featuring standards enthusiasts, vendors and implementors doing hands-on standards development (FHIR) and testing. As an attendee I picked one of the tracks – bulk data. This blog is part of a series on Bulk Data Setup and Testing based on my experience at HL7 FHIR Connectathon 29.
This article walks folks through the process of using the IBM FHIR Server’s helm chart with Docker Desktop Kubernetes and getting it up and running.
The upper limits for prepared statement parameters in db2, postgres and derby.
As of IBM FHIR Server 4.10.2… A colleague of mine is entering into the depths of the IBM FHIR Server’s Bulk Data feature. Each tenant in the IBM FHIR Server may specify multiple storageProviders. The default tenant is assumed, unless specified with the Http Headers X-FHIR-BULKDATA-PROVIDER and X-FHIR-BULKDATA-PROVIDER-OUTCOME. Each tenant’s configuration may mix the different providers, however each provider is only of a single type. For instance, minio is aws-s3 and default is file and az is azure-blob. Note, type http is only applicable to $import operations. Export is only supported with s3, azure-blob and file. File Storage Provider Configuration…
My team uses GitHub Actions 18 in total jobs across about 12 workflows. When we get multiple pull requests we end up driving contention on the workflows and resources we use. I ran across concurrency control for the workflows. To take advantage of concurrency control add this snippet to the bottom of your pull request workflow: When you stack the commits you end up with this warning, and the prior job is stopped:
Tracing the IBM FHIR Server file access on MacOSX
Thie attached GIST is a package of Kubernetes yaml files and Java code to test locally with Docker/Kubernetes with the IBM FHIR Server. You’ll want to kubectl apply -f <filename> for each of the files. Then apply the fhir-server-config-snippet.json to your fhir-server-config.json And run kubectl config use-context docker-desktopkubectl -n fhir-cicd-ns port-forward kafka-0 9092 Thanks to https://github.com/d1egoaz/minikube-kafka-cluster for the inspiration.
HL7 FHIR® Da Vinci Health Record Exchange (HREX) is an FHIR Implementation Guide at version 0.2.0 – STU R1 – 2nd ballot. The HREX Implementation Guide is a foundational guide for all of the Da Vinci guides which support US payer, provider, member and HIPAA covered entity data exchange. The guide defines “FHIR profiles, operations” and depends on HL7 FHIR® US Core Implementation Guide STU3 Release 3.1.0. In an issue, I implemented this profile and operation.