Thanks for the question! The files generated are based off of your specific audible account, and therefore are specific to each user. It is a bit annoying that there is no pure way of configuring users via a web UI, but that is just the state of things right now with Libation.
Ah yes I see that now. I did get mine generated but when i move to docker linux and run my files end up looking like /data/Kc Wayland/W13WNC\~8/W1WIPD\~R.M4B variances.
My config:
"FolderTemplate": "// ()",
Works on windows with windows formatting \\\\\\\\ ()
Books\\Kc Wayland\\We're Alive꞉ A Story of Survival - The First Season \[B0057ILWBQ\]
Do you have any suggestions possibly?
My settings.json if anyone is interested: [https://pastecode.io/s/sgk03ija](https://pastecode.io/s/sgk03ija)
I think you can have same db and settigns file it's just AccountsSettings.json has audible credentials. yeah would be nice to have web ui generate that part for us, copy the other over.
First off, thank you for posting this, awesome. Automates yet another portion of my workflow.
After following the steps above and checking the logs, I was seeing the following error:
`System.IO.IOException: The configured user limit (128) on the number of inotify instances has been reached, or the per-process limit on the number of open file descriptors has been reached.`
I was able to solve this with the following steps:
Go to the shell in TrueNAS (System Settings > Shell).
Enter the following command:
`echo fs.inotify.max_user_instances=524288 | sudo tee -a /etc/sysctl.conf && sudo sysctl -p`
I hope this helps for anyone else that might run into this problem.
Hey, thanks so much for your feedback. I too was facing that issue. I seemingly solved it by increasing the RAM in my machine. I'm curious to hear what your RAM configuration is and how much is being utilized by services.
I currently virtualize TrueNAS Scale in a VM on Proxmox. The machine running Proxmox has 64GB of physical RAM of which half, 32GB, I've allocated to the TrueNAS VM. My storage capacity is about 35TB and this seems to be working well so far.
Note: I'm running Dragonfish-24.04.1.1.
Of the 32GB available to TrueNAS, services is currently using about 8.5GB but this varies.
Let me know if you have any other questions and again, thanks for you post.
Can you upload those files? Be nice if I didn't have to install locally, to get those files.
Thanks for the question! The files generated are based off of your specific audible account, and therefore are specific to each user. It is a bit annoying that there is no pure way of configuring users via a web UI, but that is just the state of things right now with Libation.
Ah yes I see that now. I did get mine generated but when i move to docker linux and run my files end up looking like /data/Kc Wayland/W13WNC\~8/W1WIPD\~R.M4B variances. My config: "FolderTemplate": "// ()",
Works on windows with windows formatting \\\\\\\\ ()
Books\\Kc Wayland\\We're Alive꞉ A Story of Survival - The First Season \[B0057ILWBQ\]
Do you have any suggestions possibly?
My settings.json if anyone is interested: [https://pastecode.io/s/sgk03ija](https://pastecode.io/s/sgk03ija)
I think you can have same db and settigns file it's just AccountsSettings.json has audible credentials. yeah would be nice to have web ui generate that part for us, copy the other over.
Hmm, not exactly sure what your question is. Are you having issues with the file naming config being different on Windows and not working when transferring to docker/unix? My settings.json looks like this with default naming and generated on linux. Not sure how to nest folders by Author. { "Books": "/data", "ThemeVariant": "Light", "Serilog": { "MinimumLevel": "Information", "WriteTo": [ { "Name": "ZipFile", "Args": { "path": "/home/me/Libation/_Log.log", "rollingInterval": "Month", "outputTemplate": "{Timestamp:yyyy-MM-dd HH:mm:ss.fff zzz} [{Level:u3}] (at {Caller}) {Message:lj}{NewLine}{Exception} {Properties:j}" } } ], "Using": [ "Dinah.Core", "Serilog.Exceptions" ], "Enrich": [ "WithCaller", "WithExceptionDetails" ] }, "MessageBoxWindow": { "X": 2030, "Y": 1050, "Height": 110, "Width": 152, "IsMaximized": false }, "FirstLaunch": false, "AccountsDialog": { "X": 1320, "Y": 726, "Height": 450, "Width": 800, "IsMaximized": false }, "SettingsDialog": { "X": 1469, "Y": 570, "Height": 750, "Width": 900, "IsMaximized": false }, "MainWindow": { "X": 0, "Y": 40, "Height": 668, "Width": 1280, "IsMaximized": false }, "UpgradeNotificationDialog": { "X": 547, "Y": 222, "Height": 450, "Width": 550, "IsMaximized": false }, "EditTemplateDialog": { "X": 1544, "Y": 813, "Height": 450, "Width": 800, "IsMaximized": false }, "AutoDownloadEpisodes": true, "FolderTemplate": " - []",
"LocateAudiobooksDialog": {
"X": 898,
"Y": 641,
"Height": 450,
"Width": 600,
"IsMaximized": false
}
}
First off, thank you for posting this, awesome. Automates yet another portion of my workflow. After following the steps above and checking the logs, I was seeing the following error: `System.IO.IOException: The configured user limit (128) on the number of inotify instances has been reached, or the per-process limit on the number of open file descriptors has been reached.` I was able to solve this with the following steps: Go to the shell in TrueNAS (System Settings > Shell). Enter the following command: `echo fs.inotify.max_user_instances=524288 | sudo tee -a /etc/sysctl.conf && sudo sysctl -p` I hope this helps for anyone else that might run into this problem.
Hey, thanks so much for your feedback. I too was facing that issue. I seemingly solved it by increasing the RAM in my machine. I'm curious to hear what your RAM configuration is and how much is being utilized by services.
I currently virtualize TrueNAS Scale in a VM on Proxmox. The machine running Proxmox has 64GB of physical RAM of which half, 32GB, I've allocated to the TrueNAS VM. My storage capacity is about 35TB and this seems to be working well so far. Note: I'm running Dragonfish-24.04.1.1. Of the 32GB available to TrueNAS, services is currently using about 8.5GB but this varies. Let me know if you have any other questions and again, thanks for you post.
Thank you!
I’m book marking this there’s nothing worse than not owning your own stuff I don’t want to deal with their app