There are some instances where you may wish to package a set of processes that ready for installation by managed accounts. Dell Boomi offers a solution to this problem with the use of Integration packs. Integrations packs allow you to offer a pre-packaged solution to your end users.
There are however a few things to consider before using integration packs:
You can only have a maximum of 100 processes in a single integration pack.
There are several components that cannot be added to an integration pack. Those elements included
Process Route Components
Processes that contain Process Route Components
Processes must first be published to a process library before being added to an integration pack. For more information about how to use process libraries in Dell Boomi, see this link
Integration pack management and multi-install integration packs are optional Dell Boomi AtomSphere features that must first be…
A common operation in Dell Boomi is to copy processes from one account to another. These processes can often be used as templates for new processes. One way to accomplish this task is to copy the process to the other account choosing the account and the folder.
While this method does move the process that we need to the correct account, there are a few drawbacks to this method.
It is difficult to determine which version of the process you are using when copying the process in this manner.
Copying a new version of the process is often messy process which involves first deleting the old version of the process before copying over the new version.
In many instances there are better and less painful alternatives to the above mentioned method.
Dell Boomi offers a feature called Process Library Management that fixes these issues….
While the Dell Boomi user interface is highly robust and user friendly there are times when using the user interface can prove to be quite tedious. One example that we have discovered to be an example of this are process deployments. In order to deploy a set of Boomi processes one must search for each individual process and deploy. This may not seem overly tedious if AtomSphere only has a few processes. But in some cases AtomSphere could have numerous applications each consisting of many processes. It is easy to see given the scenario how frequent deployments could prove to be very time consuming.
Luckily, there is an alternative to using the AtomSphere user interface. Dell Boomi offers a set of APIs (both REST and SOAP) that can be used for various sets of tasks and that includes deploying Boomi processes….
If there’s one problem that an integration system needs to address, it’s standardizing the format of all data being exchanged. Before any type of data can be processed or transformed, the system handling it must be aware of its format and structure. Using a service, at the client endpoint, forces the user to provide proper data that is agreed upon. This acts as a contract between the service and the client. Given a certain format of data, the client will receive another format of data in return. Anything that happens behind the curtain is the responsibility of the service; the client simply waits for the returned data set.
BizTalk WCF Service Publishing Wizard
One of the more popular tools offered by BizTalk 2010 is the WCF Service Publishing Wizard. This handy tool can automatically create a WCF service, and fully configures the…
The Dell Boomi AtomSphere provides users with the ability to create and configure trading partners for EDI transactions. Users can create trading partners to be used with the X12 Interchange format.
The X12 Interchange Format is a standard used for the sending and receiving of EDI Software files, it allows us to define different pieces of information about the file and communicate information about what is expected of the file between the respective trading partners. It uses the ISA and GS segment definition headers within sent EDI messages to set this information.
This post will cover how to create a basic X12 Trading Partner and the different configuration options available for them.
First we can create our trading partner by creating a new component and selecting “Trading Partner”, we can configure items such as the name and whether or not this is our company or…
Through the use of Atom Queues and Listeners a single process can spawn many iterations of a listening process. Each process spawned from a listener will execute asynchronously, independent of any other executions. By default, an Atom Queue listener will spawn an instance of the listening process every time a document is written to the queue. While this will work fine in cases with a low throughput, larger numbers of documents being processed will cause a large number of executions to get kicked off.
Many times when working with Microsoft Azure, it is necessary to read and write information to Microsoft Azure File Storage. Unfortunately, at this time Dell Boomi currently only supports a Microsoft Blob Storage Connector; one possible solution to this problem is to use Microsoft’s AzCopy utility to read and write information to Microsoft Azure File Storage.
AzCopy is a command-line Utility designed for copy data to and from Microsoft Azure Blob, File, and Table storage; but our purposes we will be using it to write data to Microsoft Azure File Storage.
AzCopy can be downloaded from this link (http://aka.ms/downloadazcopy). To install, simply follow the installation instructions.
One possible use of AzCopy is to upload newly created data files to Microsoft Azure File Storage. By doing so we can call AzCopy to upload the data files after we have finished creating them all rather…
Many times when a document cache is created for use in a lookup function, only a few values are actually needed. A typical approach to document caching would write an entire document to the cache, including unneeded values. This approach works fine when caching a small number of documents, but will perform worse as more documents need to be cached. A much more efficient approach is to only cache the elements of a document that are actually used.
Once a process is deployed, any debugging or error tracing is typically done through the process logs. Although this method of debugging answers where in the process and error occurred, the developer lacks the answer to a crucial debugging question: “What data caused this error?” In test mode, a developer can simply look at the Shape Source Data of the component where the error occurred. After deployment, that level of granular debugging is not present. So how can we see the documents that were passed through the process?
What is an API?
What exactly is an API? We hear about how valuable APIs are all the time; MuleSoft founder, Ross Mason, often says that they will have a significant impact on business. However, what is the definition of an API?
API stands for Application Programming Interface. An API is a software intermediary that allows two applications to talk to each other. In other words, an API is the messenger that delivers your request to the provider that you’re requesting it from and then delivers the response back to you.
An API defines functionalities that are independent of their respective implementations, which allows those implementations and definitions to vary without compromising each other. Therefore, a good API makes it easier to develop a program by providing the building blocks.
When developers create code, they don’t often start from scratch. Fundamental to developer productivity is…