Upgrading SQL server to SP4 or SQL Server 2016 encountered error 'msdb110_upgrade.sql' encountered error 2714, state 6, severity 25
When doing an dummy run for a SQL instance upgrade, I encountered this error and it resulted in the SQL service not starting after upgrade and the upgrade wizard reporting errors.
After a couple of attempts I had to dig into it to find what was going on so referencing the article SQL Server Service fails to start after applying patch. Error: CREATE SCHEMA failed due to previous errors I tried deleting the DatabaseMailUserRole Schema from msdb but the server still failed to start.
This was SQL Server 2012 so I checked the SQL server logs found in at C:\Program Files\Microsoft SQL Server\MSSQL10_50.MSSQLSERVER\MSSQL\Log\ERRORLOG.txt (that path may vary between versions).
The tail of the log looked like the following, where I had already removed the DatabaseMailUserRole, so it may have reported that too previously:
019-02-15 17:04:20.53 spid4s Setting object permissions...
2019-02-15 17:04:20.63 spid4s Error: 2714, Severity: 16, State: 6.
2019-02-15 17:04:20.63 spid4s There is already an object named 'TargetServersRole' in the database.
2019-02-15 17:04:20.63 spid4s Error: 2759, Severity: 16, State: 0.
2019-02-15 17:04:20.63 spid4s CREATE SCHEMA failed due to previous errors.
2019-02-15 17:04:20.63 spid4s Error: 912, Severity: 21, State: 2.
2019-02-15 17:04:20.63 spid4s Script level upgrade for database 'master' failed because upgrade step 'msdb110_upgrade.sql' encountered error 2714, state 6, severity 25. This is a serious error condition which might interfere with regular operation and the database will be taken offline. If the error happened during upgrade of the 'master' database, it will prevent the entire SQL Server instance from starting. Examine the previous errorlog entries for errors, take the appropriate corrective actions and re-start the database so that the script upgrade steps run to completion.
2019-02-15 17:04:20.63 spid4s Error: 3417, Severity: 21, State: 3.
2019-02-15 17:04:20.63 spid4s Cannot recover the master database. SQL Server is unable to run. Restore master from a full backup, repair it, or rebuild it. For more information about how to rebuild the master database, see SQL Server Books Online.
2019-02-15 17:04:20.63 spid4s SQL Server shutdown has been initiated
2019-02-15 17:04:20.63 spid4s SQL Trace was stopped due to server shutdown. Trace ID = '1'. This is an informational message only; no user action is required.
2019-02-15 17:04:20.80 spid14s The SQL Server Network Interface library successfully deregistered the Service Principal Name (SPN) [ MSSQLSvc/***] for the SQL Server service.
2019-02-15 17:04:20.80 spid14s The SQL Server Network Interface library successfully deregistered the Service Principal Name (SPN) [ MSSQLSvc/***:1433 ] for the SQL Server service.
So it seems I had the same problem as the referenced blog post but with another schema role.
After running the upgrade and it failing, I started SQL server with the trace flag 902, using net start mssqlserver /T902 from an elevated command prompt. This prevents the startup scripts running.
Then connected to the SQL server instance using SSMS and located the 'TargetServersRole' under the msdb database and right click, deleted it.
I then stopped SQL server and restarted it normally, without the trace flag, so it runs the 'msdb110_upgrade.sql' at startup again.
This time it started normally.
If this was helpful please comment as it motivates me to blog more!
CREATE VIEW myschema.Admin_RunningSQLJobs
,DATEDIFF(SECOND, activity.run_requested_date, GETDATE()) AS Elapsed
FROM msdb.dbo.sysjobs_view job
JOIN msdb.dbo.sysjobactivity activity ON job.job_id = activity.job_id
JOIN msdb.dbo.syssessions sess ON sess.session_id = activity.session_id
SELECT MAX(agent_start_date) AS max_agent_start_date
) sess_max ON sess.agent_start_date = sess_max.max_agent_start_date
WHERE run_requested_date IS NOT NULL
AND stop_execution_date IS NULL
Using this view we can check for a running job before running our job…
In this example we don’t want the nightly pricing build to run if a monthly build is in progress, note, nightly is scheduled later than monthly. So wrap the call to the stored procedure in the job step like this…
IF NOT EXISTS(SELECT * FROM myschema.Admin_RunningSQLJobs WHERE name='Pricing - Monthly Rebuild')
EXEC myschema.Nightly Pricing build
Note about backups:
We restore the live company into the test company of our ERP system, once a week, to keep the test company data fresh. It is worth noting that it is wise to keep the above SQL view in a database that is not involved in a restore, otherwise errors will occur trying to access the view as that database is in restore mode and cannot be accessed by the SQL job.
For this reason we keep this view in an “admin” database that is not part of the automated restore process, using the view in SQL jobs used against the test database prevents them erroring when the restore process in in progress.
In previous blog posts we have talked about the way that other applications can drill down into parts of GP by using an direct approach via a pluggable protocol handler that is installed when GP is installed. Sometimes this does not work, to debug if the protocol handler has an end point to talk to follow these instructions.
Download Handle.exe which is part of the SysInternals tools that Microsoft acquired some time back, written by Mark Russinovich,
Extract the Zip file, if you have 64 bit system use the 64 bit version, otherwise use the plain version.
Use PowerShell to pipe out and decode handle.exe output
Open a PowerShell editor on your machine (e.g. start menu, search powershell, launch Windows PowerShell ISE)
In my case I extracted the handles.exe into the C:\Users\tw\Documents\ folder, you will need to change the path for where you extracted it to…
Execute the following PowerShell Command:
C:\Users\tw\Documents\handle64.exe net.pipe -accepteula
This will accept the user agreement and pipe the output of the handle viewer to powershell for decoding.
This gives the following output.
You can see that visual studio and Dynamics.exe have named pipes running.
If I close Dynamics GP, then run it again…
You see that Dynamics .exe is not listed as having named pipes available anymore as it is closed.
Using powershell to list all
The PowerShell modules for GP allow a number of DevOps activities. For example you may install the Dynamics GP Lesson (sample) company TWO and install the system database via these PowerShell modules. I take full advantage of the PowerShell modules when running and building Docker images for Dynamics GP.
Today whilst doing this I encountered the following error when moving to build the image for a different version of Dynamics GP:
Macro failed to complete successfully.
The PowerShell script works by creating a macro file in the GP\data folder, the PowerShell script then runs DexUtils passing the macro file as a command parameter. The macro file tells utilities what it needs to do (in this case create the system database), you can see this macro in the folder below, captured after this error occurred in the Docker container. CreateSysetmDatabase.mac is the macro file that PowerShell is using to create the system database.
We have no idea at this point what the error actually is, as the error description of “Macro failed to complete successfully” is not very helpful. However you will notice also in that folder the “DexMacro.log” file. Lets look in that file by reading it with the command “type DexMacro.log”.
So below is the output of that command, the contents of the log.
There we can see the issue, I’m using SQL Server 2017, I already know this is only compatible with GP2018 onwards. However you will notice from the folder name I’m trying to use it with GP2016. The error is badly worded, it says you need to upgrade to server 11 from 14. It actually means you need to use an older version of SQL for this version of Dynamics GP. Something I already knew but my focus wasn’t in this area at the time I was developing the Docker container.
Please comment if you found this helpful – motivates me to blog more!