You are viewing limited content. For full access, please sign in.

Question

Question

Smart Fields Performance - Do you get your own QoS?

asked on January 6

Does each customer have their own private compute power when processing through smart fields or is everyone sharing a single pipeline where they could notice different compute performance on different days depending on global demand?

0 0

Replies

replied on January 6

Neither laugh While the infrastructure and resources are shared, there are controls in place to ensure one busy customer will not dominate the queue and slow everybody down. Scaling rules are in place to add capacity as load increases and decommission it as it decreases.

2 0
replied on January 7

QoS means a guaranteed compute where one customer can not slow down another. Does this mean that the performance will vary based on global demand to some extent. We are getting results all over the place. Some days it takes minutes and some days it takes hours to process roughly 100 invoices.

0 0
replied on January 7 Show version history

Is there a practical business impact to it taking "hours to process roughly 100 invoices" for these customers running Smart Fields though backend process automation?

Separately, that's not the commonly used meaning of Quality of Service in a technical context, which is in almost all cases about networking. To quote Wikipedia's Quality of service article directly:

In the field of computer networking and other packet-switched telecommunication networks, quality of service refers to traffic prioritization and resource reservation control mechanisms rather than the achieved service quality. Quality of service is the ability to provide different priorities to different applications, users, or data flows, or to guarantee a certain level of performance to a data flow.

Abstracted to the application layer, QoS is a prioritization strategy. It doesn't prescribe a specific technical implementation like dedicated compute resources per tenant.

The strategy we use that Miruna outlined, auto-scaling and "fair queuing", is a modern, scalable, cloud-native approach. Within those there are various configuration levers - different scaling rules, fair queuing thresholds, etc. Fair queuing is the main mechanism to directly address "noisy neighbor" impact.

I know we actively service monitor metrics like average/max job time and adjust those configuration levers to try to keep relevant performance metrics within our service targets (which I don't know the specifics of).

0 0
replied on January 7

As for the business case, I guess they are just trying to get the invoices out faster rather than waiting a couple hours.

The reason we are asking about QoS is because it is the measurement of the overall performance of a service through techniques in services that manage traffic and resources to guarantee performance. When we see massive varying in performance we can only guess that the impact is either one of the following:

  • Something about the specific documents we are processing that takes a long time to analyze
  • Someone else in the world who we are waiting in a queue behind
0 0
replied on January 8

It's hard to say without seeing the whole picture, but I think you may be jumping to conclusions. While in most circumstances processing is expected to be within minutes, there may be other factors affecting processing speed. The upper bound for processing is about 4 hours before a processing request should be considered suspect/failed. 

There are various rate limits that could come into play for a given customer even without activity from other customers. So whether it was a batch of invoices or 100 invoices trickling through the day may make a difference. Whether workflow was involved may also make a difference. The number of pages could also be a factor.

If you want to look into this further, please open a case with Laserfiche Support and indicate the customer and the timeframe when processing occurred. 

1 0
replied on January 12

An update on this from the customer. The customer who was running into the performance problems was also running into a few failed documents (although it is hard to know this because there is no failed document metadata field in the interface). They just noticed no metadata was being set.

After we opened a support ticket and Laserfiche fixed the cause of the failure (have any blank page on a document), they have not had any performance issues. Could it be that a document failing causes all the rest in the queue to get stuck?

Everything is processed in under a minute now where it was taking hours before.

0 0
replied on January 12 Show version history

(The following applies to the asynchronous calls for processing the LF Server makes when the template is applied from a background application like Workflow. It does not apply to requesting extraction directly from the web client)

Not likely, it's not a single file queue, it's a distributed architecture. So while there is one incoming queue for requests for processing, they are de-queued and dispatched for processing to a farm with multiple workers. Depending on the size of the document, processing individual documents may have different durations. If processing needs to retry due to issues in the underlying components, then the document goes into a separate retry queue. 

The end result is that while documents are sent in one order, the results may not be in the same order. The Laserfiche Server checks for completed requests periodically, pulls the data for the ones that are done and sets metadata. The LF Server periodically checks for a request's result until it receives a completed/failed status from Laserfiche Cloud. 

What your customer was seeing was the results of some documents going through retries. In their case, they were not seeing delays in all processing, just the specific documents that had blank pages. 

2 0
replied on January 13

Ok let me follow up and see if this is what was happening. I got the impression that when processing between 30 and 100 documents it would take hours each time where only a few of these documents had blank pages causing failures.

0 0
You are not allowed to follow up in this post.

Sign in to reply to this post.