Does the TCPIP datasource module support reading UDP devices? If not, what alternatives are there for reading UDP, short of writing a custom module?
Best posts made by Pedro
-
Reading a UDP device using the TCPIP datasource module
Latest posts made by Pedro
-
RE: Reloading log4j2.xml without a Mango 3.7.12 restart
@MattFox thank you for your reply. I posed the question in this forum while I was waiting for an official response from Radixiot. Now I'm posting the answer here because I finally heard an official response from Radixiot today: they stated there is a "LOG4J RESET" button somewhere in the UI.
I subsequently located that button on the following page in Mango 3.7:
Administration > Admin Home > Utilities > Log4J Reset
or: https://mangoserverurl/ui/administration/log4j-resetI would have responded to your post sooner, but I was waiting for an official answer that would make for a more useful reply, not knowing it would be 19 days. I calculated that they are receiving over 400 support tickets over a 10-day period; that's a lot to keep up with.
-
Reloading log4j2.xml without a Mango 3.7.12 restart
I read somewhere that Mango now supports reloading log4j2 files without a restart. How does one induce Mango 3.7.12 to read '/opt/mango/overrides/properties/log4j2.xml' without a restart? Can this be done from the web interface, or only subsequent to an xml file edit and Mango restart?
If Mango 3.7.12 cannot read the log4j2 file without a restart, can I add the
monitorInterval
syntax intolog4j2.xml
so that Mango can re-read it periodically without a Mango restart? What is the exact log4j2 file syntax recommended to enable re-reading the file?Thanks.
-
RE: Convert graphicalViews to Dashboard
@terrypacker thanks, but this is very unfortunate, as I am not thinking of "a much nicer page" in that I probably have 20 pages: most have dozens or hundreds of points. Converting by hand would be very tedious.
Is there a reference that explains how to recreate dashboard structures that are equivalent to existing structures used in the graphical view? For instance, the "Simple point" data structure normally shows the point value. On mouseover, a pop-up point links to the point details, and shows a pop-up default graph. Would I have to recreate that behavior in the dashboard from scratch?
Similarly, the "complex point" structure shows a table of Simple points, each with the Simple point mouseover behavior; binary points show pop-up default tables on mouseover. Would I have to write all that too, or are there equivalent data structures? Last time I looked at the dashboard a couple of years ago, I did not see a built-in way to replicate this behavior without going low-level. Even if there is an equivalent structure, mousing and clicking around to recreate each point would take ages because there are so many.
What is available to speed this conversion?
-
Convert graphicalViews to Dashboard
Is there a script that will convert Mango 3.xx graphicalViews to the equivalent Dashboard views? What is the recommended migration path? Is there a table that shows the names of equivalent data structures that exist on the two views?
-
RE: Times series database storage format, point arrays, and downsampling
@phildunlap said in Times series database storage format, point arrays, and downsampling:
quite a lot going on there!
Yes, I was considering whether to split my post into different topics. Thank you for all your answers.
The POST
/rest/v2/script/
run allows you to submit a script and get the results of its run.That sounds very interesting. I just completed a python script using XSRF token. Nice: I like how it eliminated the need to login.
One could easily downsample their data via script, which I can provide a simple or efficient example of if desired.
If you already have an existing script, it would be nice if you could post it under its own forum topic, as I'm sure many of us would like to downsample old data. In addition to downsampling old data, I have numerous points that were logged much too often due to initially setting a log tolerance threshold that was too small. However, I won't have time to run this script right away due to my other project.
there is no data type that is "array of data type: numeric" and handled in Mango as such.
You understood me correctly. I'm looking to store an array of readings (as in multiple channels for each timestamp). Basically, a 2D numerical array where the rows are for different timestamps and the columns are the same type of data type but from different sources (channels). If it were stored in a CSV or spreadsheet, it would look like this:
timestamp t+0, channel[1],channel[2],channel[3],channel[4],channel[5],...,channel[1000] timestamp t+1, channel[1],channel[2],channel[3],channel[4],channel[5],...,channel[1000] timestamp t+2, channel[1],channel[2],channel[3],channel[4],channel[5],...,channel[1000] timestamp t+3, channel[1],channel[2],channel[3],channel[4],channel[5],...,channel[1000] timestamp t+4, channel[1],channel[2],channel[3],channel[4],channel[5],...,channel[1000] ...
It seems to me that in order to reduce data redundancy (by not storing the same timestamp multiple times) I could store the data in HDF5 format. HDF5 includes the metadata for the stored information, so the data can be retrieved into a meaningful format using generic tools, even without the source code that stored it. Additionally, it can efficiently compress and decompress binary data such as numerical arrays. My array elements could be any number of bytes. HDF5 is also extremely fast.
Summary Points - Benefits of HDF5
- Self-Describing The datasets with an HDF5 file are self describing. This allows us to efficiently extract metadata without needing an additional metadata document.
- Supports Heterogeneous Data: Different types of datasets can be contained within one HDF5 file.
- Supports Large, Complex Data: HDF5 is a compressed format that is designed to support large, heterogeneous, and complex datasets.
- Supports Data Slicing: "Data slicing", or extracting portions of the dataset as needed for analysis, means large files don't need to be completely read into the computers memory or RAM.
- Open Format - wide support in the many tools: Because the HDF5 format is open, it is supported by a host of programming languages and tools, including open source languages like R and Python and open GIS tools like QGIS.
I also found TsTables, a PyTables wrapper that will enable storing time stamped arrays into HDF5 files in daily shards and seamlessly stitch them together during queries. Appends are also efficient. The HDF5 tools will also help for debugging whether any inconsistency is occurring during the read or write operation.
-
Times series database storage format, point arrays, and downsampling
I'm looking into reading over 2,000 new datapoints, each reporting once per minute, and storing them either in Mango or in another time-series database. Most values are 16 bits wide, but many may be only 8 bits wide, and all could be stored with the same timestamp each minute. This is a lot of data, so I'm looking to store it efficiently, and then probably analyze and visualize it with the Anaconda Python toolset.
I currently have almost 1,300 enabled points in a Mango installation. I recommend the command line tool
jfq
to useJSONata
queries to quickly answer such questions:$ jfq '$count(dataPoints[enabled=true].[name])' Mango-Configuration-Aug-01-2019_000500.json
1289
The.[name]
is not needed, but I like to list the point names as a sanity check before invoking the count.This brings me to Mango's NoSQL time-series database:
- How does Mango's time-series database store binary or numerical values?
- Are all point values stored as 64 bits? Even binary values?
- Is each TSDB entry 128 bits? (64bit timestamp+64bit value)?
Each four 16-bit readings would be transmitted packed into 64-bit wide words, with the same timestamp from each data source, either over MQTT-SN into a broker and into the Mango MQTT client subscriptions, or over CoAP and REST PUTs. Those four stored numbers can then be used to generate the additional two numbers I need for each timestamp; incidentally, if you ever integrate an MQTT broker into Mango, please be sure it supports MQTT-SN like the Mosquitto broker. Values can therefore be calculated upon retrieval; storing such values would be redundant and bloat storage needs by another 50%.
To summarize, I could pack four 16-bit values into a 64-bit pointvalue, and then generate the remaining two points when I read and unpack the pointvalue, thus generating my desired six 16-bit values. Unpacking could almost be done with metadata functions, but metadata functions are triggered when the data enters Mango rather than when the data is retrieved from Mango's TSDB.
- Have you considered implementing ad-hoc metadata functions, where the data is generated when it is retrieved rather than when it is stored? This would reduce the need to store data that can simply be calculated from other stored points. Depending on the Ad-hoc metadata
Logging properties
, calculated points could either be stored or recalculated each time they're retrieved. Metadata point value calculations are performed on retrieved stored data now by clicking on the metadata point sourceGenerate history
function, but they're not ad-hoc in that they are not automatically generated when the metadata point is queried for data that is does not have.
The most efficient storage in my case would be where each 2,000 element row begins with a single timestamp, and each column is a different data value. Since the timestamp will be the same for all the points, that timestamp could be stored only once. Rare missing values could be stored with a special value like a NaN. This avoids the need to write 2,000 identical timestamps to disk. Calculating the additional 2/4 points upon value retrieval from disk would further reduce the data storage demands by 1/3.
- Is there any way to handle point arrays in Mango, where each array index is a different point? e.g. voltage[0], voltage[1], ... voltage[1000]?
- As the data ages, it could be down-sampled (decimated) from 1-minute intervals to 10-minute intervals. Most time-series databases support automatic downsampling. Are there any plans to add automated downsampling to Mango? Purging should not be the only option for old point data.
I've considered using influxDB. Although they have downsampling policies, it seems that they do not support arrays; the closest data element they support is a 64-bit number. Consequently, I would still have to pack and unpack my 16-bit values. One consequence of having to pack and unpack values is that it makes it unlikely that the visualization tools would render the stored data directly. This forces me to choose between direct visualization and storage efficiency.
Thank you for your input.
-
RE: Internal data source attribute 'Point values to be written' keeps climbing. Mitigation strategies?
As predicted, the
Point values to be written
has continued to be fairly flat since yesterday's configuration change. It is fluctuating between 276,180 values and 276,230 values, except for a brief spike to 276,300 values. It is not trending up or down. I wonder what will happen next time Igenerate history
on a Metadata point. -
RE: Internal data source attribute 'Point values to be written' keeps climbing. Mitigation strategies?
@phildunlap said in Internal data source attribute 'Point values to be written' keeps climbing. Mitigation strategies?:
the best way to zero it out would be restarting Mango.
Ironically, restarting Mango will result in a prolonged data outage while the values are being written. Therefore it becomes a matter of choosing which data I want to lose (today's, or the accumulated values). I also wonder if those values are from a particular point, or from random points.
-
RE: Internal data source attribute 'Point values to be written' keeps climbing. Mitigation strategies?
Thanks very much: 3.5 hours after changing the TSDB settings, I see that the
Point values to be written
has leveled off, on average. Plotted on a one day scale, it will surely look like a flat horizontal line. Now it would be nice to figure out how to make it go down from 276,196 values to zero.My current configuration:
$ jfq 'systemSettings' Mango-Configuration-Jul-17-2019_201152.json | grep -i nosql "mangoNoSql.writeBehind.statusProviderPeriodMs": 5000, "mangoNoSql.writeBehind.maxInstances": 10, "mangoNoSql.backupHour": 4, "mangoNoSql.backupEnabled": true, "mangoNoSql.backupMinute": 0, "mangoNoSql.writeBehind.maxInsertsPerPoint": 10000, "mangoNoSql.backupPeriods": 1, "mangoNoSql.backupFileCount": 3, "mangoNoSql.backupIncremental": false, "mangoNoSql.backupFileLocation": "/mnt/WesternDigitalUSB/mango-backup", "mangoNoSql.corruptionScanThreadCount": 100, "systemEventAlarmLevel.NOSQL_DATA_LOST": "CRITICAL", "mangoNoSql.writeBehind.minInsertsPerPoint": 1000, "mangoNoSql.writeBehind.stalePointDataPeriod": 60000, "mangoNoSql.writeBehind.batchProcessPeriodMs": 500, "mangoNoSql.writeBehind.maxRowsPerInstance": 100000, "mangoNoSql.backupLastSuccessfulRun": "Jul-13-2019_040000", "mangoNoSql.writeBehind.backdateDelay": 4985, "action.noSqlBackup": "", "mangoNoSql.writeBehind.stalePointCleanInterval": 60000, "action.noSqlRestore": "", "mangoNoSql.writeBehind.minDataFlushIntervalMs": 100, "mangoNoSql.backupPeriodType": "WEEKS", "mangoNoSql.writeBehind.spawnThreshold": 100000, "mangoNoSql.intraShardPurge": false
-
RE: Internal data source attribute 'Point values to be written' keeps climbing. Mitigation strategies?
@terrypacker said in Internal data source attribute 'Point values to be written' keeps climbing. Mitigation strategies?:
Are you seeing any Data Lost events?
I have not seen Data Lost events. However, since I've had to kill Mango to complete the last several restarts, if there were a data loss it would not have generated an event at that time.
If so these should enlighten us as to why this is happening, they get raised whenever a batch fails to write.
Thanks, that's reassuring. I had not remembered that there was such an event.
I would make sure that event level is set to something other than Do No Log too.
I had apparently set that event to critical. I did not see any such events, and in the last month I re-enabled critical events emails forwarding through my mobile phone network's email to SMS gateway. My phone sounds an attention getting submarine dive alarm when receiving Mango text messages.
Do you have any Alphanumeric points that would be saving very large strings of text?
Only one. I tried listing ALPHANUMERIC points in the table on the data_sources page, but for some reason it would not show a full query response (there were hourglasses at the bottom of the list). In any case, the JSONata command-line utility jfq enables me to specify more detailed reporting of the Mango configuration.
Query:
# Show name and loggingType and context update of all enabled alphanumeric points that are configured to log their values: $ jfq 'dataPoints[pointLocator.dataType="ALPHANUMERIC"][enabled=true][loggingType!="NONE"].[name,enabled,loggingType,pointLocator.context]' Mango-Configuration-Jul-17-2019_000500.json
I generate a large JSON structure to predict the tides based on a pressure sensor data. The table is displayed in a Mango 2.x graphical view via a server-side script graphical object that converts the JSON to an HTML table. The table is apparently using the Alphanumeric_Default template, which saves "When point value changes." However, it only updates context when the tide direction changes, which is only a handful of times per day. The only other alphanumeric point being logged has a context that triggers only twice a year.
The NoSQL database has a limit to the size of each entry, which is large but its possible this could cause these symptoms.
The
Point values to be written
rises by about 1,700 values per day, which far exceeds the number of times the tide changes direction in one day.I see there is now a Failed Login event. That's wonderful, thanks.
I just changed
the point clean interval (ms) and the stale data period (ms) to 60000
I will report back whether or not I see a change in the trend of the
Point values to be written
. It should take an hour or two to see a change.Thanks for your help.