mdsd constantly restarts in Service Fabric
See original GitHub issueI’m seeing mdsd constantly restarting in a Service Fabric deployment; it’s happening roughly twice a second. The configuration is a Service Fabric application on Azure with the stock Ubuntu 16.04 image on the scaleset. Here’s an excerpt from /var/log/syslog:
Dec 6 02:34:18 isoapi00000W systemd[1]: mdsd.service: Service hold-off time over, scheduling restart.
Dec 6 02:34:18 isoapi00000W systemd[1]: Stopped Azure MDS daemon (on systemd).
Dec 6 02:34:18 isoapi00000W systemd[1]: Starting Azure MDS daemon (on systemd)...
Dec 6 02:34:18 isoapi00000W mdsd[4086]: * Starting Azure MDS Daemon:
Dec 6 02:34:18 isoapi00000W mdsd[4086]: ...done.
Dec 6 02:34:18 isoapi00000W systemd[1]: Started Azure MDS daemon (on systemd).
Dec 6 02:34:18 isoapi00000W systemd[1]: mdsd.service: Main process exited, code=exited, status=1/FAILURE
Dec 6 02:34:18 isoapi00000W mdsd[4145]: * Stopping Azure MDS Daemon:
Dec 6 02:34:18 isoapi00000W mdsd[4145]: No mdsd found running; none killed.
Dec 6 02:34:18 isoapi00000W mdsd[4145]: ...fail!
Dec 6 02:34:18 isoapi00000W systemd[1]: mdsd.service: Control process exited, code=exited status=1
Dec 6 02:34:18 isoapi00000W systemd[1]: mdsd.service: Unit entered failed state.
Dec 6 02:34:18 isoapi00000W systemd[1]: mdsd.service: Failed with result 'exit-code'.
Dec 6 02:34:21 isoapi00000W systemd[1]: mdsd.service: Service hold-off time over, scheduling restart.
Dec 6 02:34:21 isoapi00000W systemd[1]: Stopped Azure MDS daemon (on systemd).
Dec 6 02:34:21 isoapi00000W systemd[1]: Starting Azure MDS daemon (on systemd)...
Dec 6 02:34:21 isoapi00000W mdsd[4170]: * Starting Azure MDS Daemon:
Dec 6 02:34:21 isoapi00000W mdsd[4170]: ...done.
Dec 6 02:34:21 isoapi00000W systemd[1]: Started Azure MDS daemon (on systemd).
Dec 6 02:34:21 isoapi00000W systemd[1]: mdsd.service: Main process exited, code=exited, status=1/FAILURE
Dec 6 02:34:21 isoapi00000W mdsd[4231]: * Stopping Azure MDS Daemon:
Dec 6 02:34:21 isoapi00000W mdsd[4231]: No mdsd found running; none killed.
Dec 6 02:34:21 isoapi00000W mdsd[4231]: ...fail!
Dec 6 02:34:21 isoapi00000W systemd[1]: mdsd.service: Control process exited, code=exited status=1
Dec 6 02:34:21 isoapi00000W systemd[1]: mdsd.service: Unit entered failed state.
Dec 6 02:34:21 isoapi00000W systemd[1]: mdsd.service: Failed with result 'exit-code'.
I understand that this is not terribly much to work with; what other debug information would be useful to track down the root cause and get a fix?
Issue Analytics
- State:
- Created 5 years ago
- Reactions:1
- Comments:9 (1 by maintainers)
Top Results From Across the Web
Restart-ServiceFabricPartition (ServiceFabric) | Microsoft Learn
Restarts replicas of a Service Fabric partition to simulate a data center blackout or cluster blackout scenario. In this article. Syntax; Description; Examples ......
Read more >Service Fabric Restart Application - azure powershell
I have a service fabric app that reads from KeyVault on startup. When we change the KeyVault value we have to restart the...
Read more >Permanent Corrosion Protection Solution for Carbon Steel
EonCoat is a spray-applied carbon steel treatment that completely stops corrosion. The permanent Corrosion Protection for carbon steel assets.
Read more >Sun meadow farm westies - Sport Castellano Reports
*#*#426#*#* - Debug information for Google Play service. ... BREEDING SINCE 2017 Premium quality farm theme fabric.
Read more >moon name generator - Ramozzi & Friends
Japanese Name Generator is not a professional Japanese name translator service, ... a suitable radioactive material into electricity by the Seebeck effect.
Read more >
Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free
Top Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
@johncrim did you figure out what is going on here? I have the same issue and it just keeps filling up the logs.
RE the original bug, we’re seeing the same problem on Ubuntu 18 too, even on a basic SF cluster created from the Azure portal. I had to create a new bare-bones SF cluster in Azure earlier today (to test a few things), and saw some azsec mdsd failures repeating every few seconds in the syslog (and I hadn’t run my workaround to remove those packages).
I didn’t evaluate whether these failures are the same as before, but they seemed like failures to connect and send info.