[statnet_help] Upgrading from ERGM v3.10 to v4.6

Steven M. Goodreau via statnet_help statnet_help at u.washington.edu
Sun Nov 17 18:09:29 PST 2024


Sorry, statnet_help listserv members - I meant to send this to the
statnet development team listserv, not the users listserv - my apologies!!

Steve

On 11/17/2024 9:05 PM, Steven M. Goodreau wrote:

>

> Hi all,

>

> So I was working through this email chain re Aditya's issue when I

> heard from Sam, and then spent time exploring both, given that there

> is some conceptual overlap.  You should have just seen the email re

> Sam's issues.

>

> Re Aditya's: here are my various thoughts. I don't actually have

> advice on how to improve your model further, Aditya, sorry -- I'll

> have to leave that to Carter and Pavel. This is about the issues

> related to the process getting to this point. The email chain is long,

> so if I missed anything that answers one of these questions, forgive me:

>

> - Aditya: it seems you've spent a long time trying to fit this model

> on ergm 4.x  If I understand correctly, you were able to fit it in

> ergm 3.x.  In the interest of moving your research forward, is there

> any reason you can't use the fit from 3.x for your simulations? I

> understand that ideally you would want it to fit just as easily in the

> latest version, but are you actually stuck in your research?

>

> - Carter: you had mentioned somewhere about how complex models on

> large networks sometimes just take a really long time to fit, and

> there's no way around it. Totally agree.  I think the issue, IIUC, is

> that in ergm 3.x, it didn't take nearly that long to converge. So that

> certainly suggests that it's possible to do it. Aditya, do I have that

> right?

>

> All: I understand that ergm 4.x's defaults are trying to tackle a wide

> range of models adaptively.  And that might mean that any given model

> might not be as smooth as it was in ergm 3.x, while ergm 4.x still

> represents a step forward overall.  That said, it seems as if Philip

> Leifeld had issues with a model fitting with 3.x defaults and not 4.x,

> and the advice that Chad gave then is similar to what Carter did now. 

> And it's similar to what also seemed to help somewhat with Sam's

> model.  So there does seem to be a large space of useful models that

> worked more smoothly in 3 than 4. So now I'm wondering: is there a way

> to set the parameters in ergm 4.x to create a fitting algorithm that

> perfectly matches what were the defaults in 3.x?  If so, does a list

> of the parameters and their values needed to do this exist somewhere? 

> Is it worth adding in a way to easily make that happen with a single

> argument (e.g. ergm3.defaults = TRUE)? Ideally one could set that and

> then still tweak on top of that, so that one could re-create a given

> model from ergm 3 that was mostly defaults but a few not.

>

> Thanks,

>

> Steve

>

>

> On 10/17/2024 9:52 PM, Khanna, Aditya via statnet_help wrote:

>

>> Hi Carter and Pavel,

>>

>>

>> Thank you so much for your helpful suggestions. Following your

>> feedback (and Carter’s numbering scheme below), I report the following:

>>

>>

>> 1.

>>

>> The GOF plot is here

>> <https://urldefense.com/v3/__https://github.com/hepcep/net-ergm-v4plus/blob/rhel9-setup/fit-ergms/out/oct12-2024-int1e6-samp1e6-hotelling_gof_plot.pdf__;!!K-Hz7m0Vt54!kOSN_RnPt-IoyMJSQB6xSVojkJkJZ6vGkt7f2MFWrHJBTJ1q5dVw5loF-a-USLQosgPSjxe8l_LYC0Sa3bsSYS53Fw62-fxr0023Ng$>and

>> the numerical summary is here

>> <https://urldefense.com/v3/__https://gist.github.com/khanna7/f263b14b7bbc09e704ecd99bda8f215d__;!!K-Hz7m0Vt54!kOSN_RnPt-IoyMJSQB6xSVojkJkJZ6vGkt7f2MFWrHJBTJ1q5dVw5loF-a-USLQosgPSjxe8l_LYC0Sa3bsSYS53Fw62-fxcx0wkHg$>.

>> Based on these, it seems that the model has not converged. When

>> we simulate from our best fit, even though it is not from a

>> converged model, the simulated networks seem not far off from the

>> mean target statistics.

>>

>> 2.

>>

>> We have looked into the underlying uncertainty in the target

>> statistics themselves, and these uncertainty regions are quite

>> large. My question is: what is the best approach to handle these

>> uncertain ranges in the target statistics? Is there a way to

>> specify not just the mean but a range of target values for each

>> of the specified ERGM parameters? I can conceptualize other ways,

>> for instance, sampling values from the uncertainty ranges for

>> each parameter, and fitting ERGMs to those configurations, for

>> instance. But I am not sure if there is a recommended heuristic

>> to pursue?

>>

>> 3.

>>

>> I also tried your suggestion to start with a non-empty graph

>> (whereas case 1 above is the GOF output on a fit that was

>> generated where I started from an empty graph). The GOF plot is

>> here

>> <https://urldefense.com/v3/__https://github.com/hepcep/net-ergm-v4plus/blob/rhel9-setup/fit-ergms/out/gof_plot_non-empty-net-all-plos1-mcmc-int1e6-samp1e6-hotelling.pdf__;!!K-Hz7m0Vt54!kOSN_RnPt-IoyMJSQB6xSVojkJkJZ6vGkt7f2MFWrHJBTJ1q5dVw5loF-a-USLQosgPSjxe8l_LYC0Sa3bsSYS53Fw62-fyZjY622Q$>.

>> I also combined the idea of starting with a non-empty graph with

>> Pavel’s suggestion to not specify any arguments in

>> control.ergmand let the algorithm figure it out (or just specify

>> the san in the control.ergm). That didn’t work either (see Rout

>> <https://urldefense.com/v3/__https://github.com/hepcep/net-ergm-v4plus/commit/e65d499d4a1bbf5f5cfe43e8355f062045f090ef*diff-72f130f6ce8c1ff7ced566c43e68f17d2232c69c7e3fbcd8d583e040e8b61890__;Iw!!K-Hz7m0Vt54!kOSN_RnPt-IoyMJSQB6xSVojkJkJZ6vGkt7f2MFWrHJBTJ1q5dVw5loF-a-USLQosgPSjxe8l_LYC0Sa3bsSYS53Fw62-fwRKZajUw$>from

>> the session).

>>

>>

>> Thank you in advance for any thoughts you are able to share,

>>

>> Aditya

>>

>>

>>

>> On Sat, Sep 7, 2024 at 8:28 PM Carter T. Butts <buttsc at uci.edu> wrote:

>>

>> Hi, Aditya -

>>

>> On 9/5/24 12:57 PM, Khanna, Aditya wrote:

>>>

>>> Hi Carter,

>>>

>>>

>>> Thank you so much for your helpful response as always. I have

>>> organized my report in terms of the various things you suggest.

>>>

>>>

>>> Verifying MCMC, GOF and the “second” MCMC: Yes, the ERGM for the

>>> model described below does converge, but, despite having

>>> converged, the simulated networks don’t seem to statistically

>>> capture the targets. I did make the GOF

>>> <https://urldefense.com/v3/__https://github.com/hepcep/net-ergm-v4plus/blob/rhel9-setup/fit-ergms/out/updated-with-oct12-2024-synthpop-ergmv4-6-all-plosone-terms-increase-mcmc-1e6-hotelling.pdf__;!!CzAuKJ42GuquVTTmVmPViYEvSg!LKGctG3KpUbsSQwk8O2E14uWvIjFBhbEa_mHZ5i_MJ9vtXw_UfiYpxvp4p1HYdnN3aFwRdIYqyjPTzdZdnlas3N1$>plots

>>> as well. Most of the terms look good, though some have peaks

>>> that are off from the zero. In general, however, I have come to

>>> rely more on actually simulating  networks from the fitted ERGM

>>> object (what I think you mean by “second MCMC run”) in addition

>>> to the GOF plots. Usually I consider my goal fulfilled if the

>>> simulated network objects capture the targets, even if the GOF

>>> plots don’t look perfect.

>>>

>>>

>> There seems to be some confusion here: setting aside exotica, if

>> your mean in-model statistics don't match your observed in-model

>> statistics, then your model has not converged.  The matching here

>> that matters is from an MCMC run from the fitted model, which is

>> what the gof() function does (but this is /not/ what you get from

>> MCMC diagnostics on the fitted ergm() object, which should show

>> the penultimate run - those are handy for diagnosing general MCMC

>> issues, but need not show convergence even when the final

>> coefficients yield a convergent model).  The plots you linked to

>> under "GOF" above seem to be MCMC diagnostics from an ergm()

>> object, not gof() output.  I'll come back to "exotica" below, but

>> first it is important to be sure that we are discussing the same

>> thing.

>>

>>> Model Convergence and tightening the MCMC tolerances:In terms of

>>> tightening the MCMC tolerances, I did increase the MCMC interval

>>> to 1e9, of the order of O(N^2). But this particular

>>> specification timed out after 120 hours, and I didn’t try to run

>>> it for longer time than that.

>>>

>>>

>> Unfortunately, there's no "that's longer than I want it to be"

>> card that convinces mixing to happen more quickly: if your model

>> is really going to take 1e9 or more updates per step to converge,

>> and if that takes longer than 120 hours on your hardware, then

>> that's how long it's going to take.  If there were magic sauce

>> that could guarantee great convergence with minimal computational

>> effort every time, it would already be in the code.  That said, I

>> don't know that all other options have yet been ruled out; and,

>> when folks encounter expensive models on large networks, there

>> are various approximate solutions that they may be willing to

>> live with.   But if one is working with a dependence model on

>> >=1e4 nodes, one must except that one may be in a regime in which

>> gold-standard MCMC-MLE is very expensive.  Just sayin'.

>>

>>> Alternate parameters to tighten the MCMC: I have experimented

>>> with the MCMC sample size and interval parameters, but have not

>>> been able to improve the quality of the simulated network. I am

>>> not as familiar with what options are available within the

>>> bounds of some reasonable computational cost.

>>>

>>> In summary, the problem remains that despite the ERGM

>>> convergence, the quality of the simulated networks suggests room

>>> for improvement, since the specified targets are not captured

>>> within the distribution of the simulated networks.

>>>

>>>

>> OK, let me see if I can offer some further advice, based on your

>> email and also something that came up in your exchange with Pavel:

>>

>> 1. We should be clear that, assuming no-exotica, you should be

>> assessing convergence from an MCMC run on the fitted model (as

>> produced by gof() or done manually).  So far, the plots I've seen

>> appear not to be runs from the fitted model, so I have not

>> actually seen evidence of the alleged phenomenon. Also, to be

>> clear, (absent exotica) if your simulated mean stats don't match

>> the observed stats (up to numerical and sometimes statistical

>> noise), your model hasn't converged.  A model that isn't

>> converging is not the same as a model that has converged but that

>> is inadequate, and the fixes are very different.

>>

>> 2. The exchange with Pavel led me to dig into your code a bit

>> more, and I realized that you are not fitting to an observed

>> network, but to target stats presumably based on design

>> estimation.  This could put you into the "exotica" box, because

>> it is likely that - due to errors in your estimated targets -

>> there exists no ERGM in your specified family whose expected

>> statistics exactly match the target statistics.  So long as they

>> aren't too far off, you still ought to be able to get close, but

>> hypothetically one could have a situation where someone gets an

>> unusually bad estimate for one or a small number of targets, and

>> their matches are persistently off; in this case, the issue is

>> that the MLE no longer satisfies the first moment condition

>> (expected statistics do not match the target statistics), so this

>> is no longer a valid criterion for assessing convergence.  If one

>> is willing/able to make some distributional statements about

>> one's target estimates, there are some natural relaxations of the

>> usual convergence criteria, and almost surely Pavel has written

>> them down, so I defer to him.  :-)  But anyway, /if/ your model

>> really seems not to be converging (by the criteria of (1)), and

>> /if/ you are using estimated target stats, then I would certainly

>> want to investigate the possibility that your model has actually

>> converged (and that you're just seeing measurement error in your

>> target stats) before going much further.  To write reckless words

>> (that you should read recklessly), one naive heuristic that could

>> perhaps be worth trying would be to look at the Z-scores

>> (t_o-t_s)/(s2_o^2+s2_s^2)^0.5, where t_o is the observed

>> (estimated) target, t_s is the simulated mean statistic, s2_o is

>> the standard error of your target estimator, and s2_s is the

>> standard error of the simulation mean.  (If you are e.g. using

>> Horvitz-Thompson, you can approximate s2_o using standard

>> results, and you can likewise use autocorrelation-corrected

>> approximations to s2_s.) If these are not large, then this

>> suggests that the discrepancies between the targets and the mean

>> stats are not very large compared to what you would expect from

>> the variation in your simulation outcomes and in your measurement

>> process.  This does not take into account e.g. correlations among

>> statistics, nor support constraints, but it seems like a

>> plausible starting point.  (Pavel and Martina have been working

>> with these kinds of problems a lot of late, so doubtless can

>> suggest better heuristics.)

>>

>> 3. Pavel's comments pointed to SAN, which also led me to observe

>> that you are starting by fitting to an empty graph.  I recommend

>> against that.  In principle, the annealer should get you to a

>> not-too-bad starting point, but in my own informal simulation

>> tests I have observed that this doesn't always work well if the

>> network is very large; in particular, if SAN dumps you out with a

>> starting point that is far from equilibrium, you are wasting a

>> lot of MCMC steps wandering towards the high-density region of

>> the graph space, and this can sometimes lead to poor results

>> (especially if you can't afford to run some (large k)*N^2 burn-in

>> - and recall that the default MCMC algorithm tends to preserve

>> density, so if the seed is poor in that regard, it can take a lot

>> of iterations to fix).  My suggestion is to use rgraph() to get a

>> Bernoulli graph draw from a model whose mixing characteristics

>> (and, above all, density) approximate the target, and start with

>> that.  An easy way to set the parameters is to fit a pilot ERGM

>> using only independence terms, use these construct a tie

>> probability matrix, and pass that to the tp argument of

>> rgraph().  Your case makes for a very large matrix, but it's

>> still within the range of the feasible.  (rgraph() does not use

>> adjacency matrices internally, and so long as you set the return

>> value to be an edgelist is not constrained by the sizes of

>> feasible adjacency matrices, but if you want to pass an explicit

>> tie probability matrix then obviously that puts you in the

>> adjacency matrix regime.) Anyway, it's better to use rgraph() for

>> this than an simulate() call, because it will be both faster and

>> an exact simulation (no MCMC).  A poorer approach not to bother

>> with mixing structure, and just to draw an initial state with the

>> right density (which at least reduces the risk that SAN exits

>> with a graph that is too sparse)....but you might as well put

>> your starting point as close to the right neighborhood as you

>> can.  The goal here is to help the annealer get you to a

>> high-potential graph, rather than expecting it to carry you there

>> from a remote location.  It is possible that this turns out not

>> to be a problem in your particular case, but it seems worth

>> ruling out.

>>

>> Hope that helps,

>>

>> -Carter

>>

>>

>>> Aditya

>>>

>>>

>>> On Fri, Aug 30, 2024 at 4:37 AM Carter T. Butts via statnet_help

>>> <statnet_help at u.washington.edu> wrote:

>>>

>>> Hi, Aditya -

>>>

>>> I'll be interested in Pavel's take on the convergence

>>> issues, but just to verify, you are assessing convergence

>>> based on a /second/ MCMC run, correct?  The MCMC statistics

>>> in the ergm object are from the penultimate iteration, and

>>> may thus be out of equilibrium (but this does /not/

>>> necessarily mean that the /model/ did not converge). 

>>> However, if you simulate a new set of draws from the fitted

>>> model and the mean stats do not match, /then/ you have an

>>> issue.  (This is why we now point folks to gof() for that

>>> purpose.)  It looks like your plots are from the ergm object

>>> and not from a gof() run (or other secondary simulation), so

>>> I want to verify that first.

>>>

>>> I also note that a quick glance at the plots from your more

>>> exhaustive simulation case don't seem all that far off,

>>> which could indicate either that the model did converge (and

>>> per above, we're not looking at a draw from the final

>>> model), or that it converged within the tolerances that were

>>> set, and you may need to tighten them.  But best to first

>>> know if there's a problem in the first place.

>>>

>>> Another observation is that, per my earlier email, you may

>>> need O(N^2) toggles per draw to get good performance if your

>>> model has a nontrivial level of dependence.  You are using a

>>> thinning interval of 1e6, which is in your case around

>>> 30*N.  It's possible that you've got too much dependence for

>>> that: O(N^2) here would mean some multiple of about 1e9,

>>> which is about a thousand times greater than what you're

>>> using. Really large, sparse networks sometimes /can/ be

>>> modeled well without that much thinning, but it's not a

>>> given.  Relatedly, your trace plots from the 1e6 run suggest

>>> a fair amount of autocorrelation on some statistics, which

>>> suggests a lack of efficiency. (Autocorrelation by itself

>>> isn't necessarily a problem, but it means that your

>>> effective MCMC sample size is smaller than it seems, and

>>> this can reduce the effectiveness of the MCMCMLE

>>> procedure.   The ones from the 1e6 run aren't bad enough

>>> that I would be alarmed, but if I were looking for things to

>>> tighten up and knew this could be a problem, they suggest

>>> possible room for improvement.)  So anyway, I wouldn't crank

>>> this up until verifying that it's needed, but you are still

>>> operating on the low end of computational effort (whether it

>>> seems like it or not!).

>>>

>>> Finally, I would note that for the stochastic approximation

>>> method, convergence is to some degree (and it's a bit

>>> complex) determined by how many subphases are run, and how

>>> many iterations are used per subphase.  This algorithm is

>>> due to Tom in his classic JoSS paper (but without the

>>> complement moves), which is still a good place to look for

>>> details.  It is less fancy than some more modern algorithms

>>> of its type, but is extremely hard to beat (I've tried and

>>> failed more than once!).  In any event, there are several

>>> things that can tighten that algorithm relative to its

>>> defaults, including increasing thinning, increasing the

>>> iterations per subphase, and increasing the number of

>>> subphases.  Some of these sharply increase computational

>>> cost, because e.g. the number of actual subphase iterations

>>> doubles (IIRC) at each subphase - so sometimes one benefits

>>> by increasing the phase number but greatly reducing the base

>>> number of iterations per phase.  The learning rate

>>> ("SA.initial.gain") can also matter, although I would

>>> probably avoid messing with it if the model is well-behaved

>>> (as here).  I will say that, except under exotic conditions

>>> in which I am performing Unspeakable ERGM Experiments (TM)

>>> of which we tell neither children nor grad students, I do

>>> not recall ever needing to do much with the base parameters

>>> - adjusting thinning, as needs must, has almost always done

>>> the trick.  Still, if other measures fail, tinkering with

>>> these settings can/will certainly affect convergence.

>>>

>>> I'd check on those things first, and then see if you still

>>> have a problem....

>>>

>>> Hope that helps,

>>>

>>> -Carter

>>>

>>> On 8/29/24 12:13 PM, Khanna, Aditya wrote:

>>>>

>>>> Hi Carter and All,

>>>>

>>>>

>>>> Thank you so much for the helpful guidance here. I think

>>>> following your suggestions has brought us very close to

>>>> reproducing the target statistics in the simulated

>>>> networks, but there are still some gaps.

>>>>

>>>>

>>>> Our full previous exchange is below, but to summarize:  I

>>>> have an ERGM that I fit previously with ERGM v3.10.4 on a

>>>> directed network with 32,000 nodes. The model consisted of

>>>> in- and out-degrees in addition to other terms, including a

>>>> custom distance term. In trying to reproduce this fit with

>>>> ergm v4.6, the model did not initially converge.

>>>>

>>>>

>>>> Your suggestion to try setting the main.method =

>>>> “Stochastic Approximation” considerably improved the

>>>> fitting. Specifying the convergence detection to

>>>> “Hotelling” on top of that brought us almost to simulated

>>>> networks that capture all the mean statistics. (Following

>>>> an old discussion thread

>>>> <https://urldefense.com/v3/__https://github.com/statnet/ergm/issues/346__;!!CzAuKJ42GuquVTTmVmPViYEvSg!K2TPlppMmLqkp0AMD_Fj8vqeLS9FI7fJTx0UxQMy3qt_I1hNTinFaFOclmyytc3bbAXyEkzYTJhdrHccGb4BZQZz$>on

>>>> the statnet github, I also tried setting the termination

>>>> criteria to Hummel and MCMLE.effectiveSize = NULL. I think,

>>>> for me, in practice, Hotelling worked a bit better than

>>>> Hummel though).

>>>>

>>>>

>>>> In general, I tried fitting the model with variants of this

>>>> <https://urldefense.com/v3/__https://github.com/hepcep/net-ergm-v4plus/blob/27736b2728965188ed73821e797b5ac7007b1093/fit-ergms/ergm-estimation-with-meta-data.R*L257-L296__;Iw!!CzAuKJ42GuquVTTmVmPViYEvSg!K2TPlppMmLqkp0AMD_Fj8vqeLS9FI7fJTx0UxQMy3qt_I1hNTinFaFOclmyytc3bbAXyEkzYTJhdrHccGU058cV1$>specification.

>>>> I got the best results with setting both MCMC

>>>> samplesize=1e6 and interval = 1e6 (see table below).

>>>>

>>>>

>>>> MCMC interval

>>>>

>>>>

>>>>

>>>> MCMC sample size

>>>>

>>>>

>>>>

>>>> Convergence Detection

>>>>

>>>>

>>>>

>>>> Results/Outcome

>>>>

>>>>

>>>>

>>>> Note

>>>>

>>>> 1e6

>>>>

>>>>

>>>>

>>>> 1e6

>>>>

>>>>

>>>>

>>>> Hotelling

>>>>

>>>>

>>>>

>>>> Closest agreement  between simulated and target statistics

>>>>

>>>>

>>>>

>>>> Max. Lik. fit summary and simulation Rout

>>>> <https://urldefense.com/v3/__https://github.com/hepcep/net-ergm-v4plus/commit/777bae726d29dae969f06e0d17b40ee59a01a7fc__;!!CzAuKJ42GuquVTTmVmPViYEvSg!K2TPlppMmLqkp0AMD_Fj8vqeLS9FI7fJTx0UxQMy3qt_I1hNTinFaFOclmyytc3bbAXyEkzYTJhdrHccGffZwKYQ$>

>>>>

>>>>

>>>> Violin plots

>>>> <https://urldefense.com/v3/__https://github.com/hepcep/net-ergm-v4plus/tree/rhel9-setup/fit-ergms/out__;!!CzAuKJ42GuquVTTmVmPViYEvSg!K2TPlppMmLqkp0AMD_Fj8vqeLS9FI7fJTx0UxQMy3qt_I1hNTinFaFOclmyytc3bbAXyEkzYTJhdrHccGZwugBZJ$>showing

>>>> the simulated and target statistics for each parameter

>>>>

>>>>

>>>>

>>>> But, I found that this was the closest I could get

>>>> producing simulated statistics that matched the target

>>>> statistics. In general, any further increasing or

>>>> decreasing of either the samplesize or interval did not

>>>> help generate a closer result, i.e., this looked to be some

>>>> optimum in the fit parameter space. I can provide further

>>>> details on the results of those fits, which for some

>>>> configurations didn’t converge, and if they did converge,

>>>> the goodness-of-fit was worse than what I had with setting

>>>> the MCMC interval and samplesize to 1e6. Based on your

>>>> experiences, I was wondering if this is expected?

>>>>

>>>>

>>>> For now, my main question is, are there any suggestions on

>>>> how I can further tune the fitting parameters to match my

>>>> targets more closely? I can provide specific details on the

>>>> outcomes of those fitting processes if that would be helpful.

>>>>

>>>>

>>>> Thanks for your consideration.

>>>>

>>>> Aditya

>>>>

>>>> On Thu, May 16, 2024 at 2:33 PM Carter T. Butts via

>>>> statnet_help <statnet_help at u.washington.edu> wrote:

>>>>

>>>> Hi, Aditya -

>>>>

>>>> I will defer to the mighty Pavel for the exact best

>>>> formula to reproduce 3.x fits with the latest

>>>> codebase.  (You need to switch convergence detection to

>>>> "Hotelling," and there are some other things that must

>>>> be modified.)  However, as a general matter, for

>>>> challenging models where Geyer-Thompson-Hummel has a

>>>> hard time converging (particularly on a large node

>>>> set), you may find it useful to try the stochastic

>>>> approximation method (main="Stochastic" in your control

>>>> argument will activate it).  G-T-H can (in principle)

>>>> have sharper convergence when near the solution, but in

>>>> practice SA fails more gracefully.   I would suggest

>>>> increasing your default MCMC thinning interval

>>>> (MCMC.interval), given your network size; depending on

>>>> density, extent of dependence, and other factors, you

>>>> may need O(N^2) toggles per step.  It is sometimes

>>>> possible to get away with as few as k*N (for some k in,

>>>> say, the 5-100 range), but if your model has

>>>> substantial dependence and is not exceptionally sparse

>>>> then you will probably need to be in the quadratic

>>>> regime.  One notes that it can sometimes be helpful

>>>> when getting things set up to run "pilot" fits with the

>>>> default or otherwise smaller thinning intervals, so

>>>> that you can discover if e.g. you have a data issue or

>>>> other problem before you spend the waiting time on a

>>>> high-quality model fit.

>>>>

>>>> To put in the obligatory PSA, both G-T-H and SA are

>>>> simply different strategies for computing the same

>>>> thing (the MLE, in this case), so both are fine - they

>>>> just have different engineering tradeoffs.  So use

>>>> whichever proves more effective for your model and data

>>>> set.

>>>>

>>>> Hope that helps,

>>>>

>>>> -Carter

>>>>

>>>>

>>>> On 5/16/24 7:52 AM, Khanna, Aditya via statnet_help wrote:

>>>>> Dear Statnet Dev and User Community:

>>>>>

>>>>> I have an ERGM that I fit previously with ERGM v3.10.4

>>>>> on a directed network with 32,000 nodes. The model

>>>>> included in- and out-degrees, in addition to other

>>>>> terms. The complete Rout from this fit can be seen

>>>>> here

>>>>> <https://urldefense.com/v3/__https://gist.github.com/khanna7/aefd836baf47463051439c9e72764388__;!!CzAuKJ42GuquVTTmVmPViYEvSg!KsbhvmLlx8TkLK7y2NKz59hK4-4H7KXVV7dEyUG4vcQzi4Mh7nO-9HupA7_ep2V2p9KkD_i00tcg6nDqczDwORmxHSho$>.

>>>>> I am now trying to reproduce this fit with ergm v4.6,

>>>>> but the model does not converge. (See here

>>>>> <https://urldefense.com/v3/__https://gist.github.com/khanna7/fbabdde53c79504dfeaebd215bb5ee20__;!!CzAuKJ42GuquVTTmVmPViYEvSg!KsbhvmLlx8TkLK7y2NKz59hK4-4H7KXVV7dEyUG4vcQzi4Mh7nO-9HupA7_ep2V2p9KkD_i00tcg6nDqczDwOW7y31IM$>.)

>>>>>

>>>>> I am looking for ideas on how to trouble shoot this.

>>>>> One suggestion I got was to set values for the "tuning

>>>>> parameters" in the v4.6 to their defaults from

>>>>> v3.11.4. But ERGM v4.6 has a lot more  parameters that

>>>>> can be specified, and I am not sure which ones make

>>>>> most sense to consider.

>>>>>

>>>>> I would be grateful for any suggestions on this or

>>>>> alternate ideas to try.

>>>>>

>>>>> Many thanks,

>>>>> Aditya

>>>>>

>>>>>

>>>>>

>>>>>

>>>>> --

>>>>>

>>>>> <https://urldefense.com/v3/__https://sph.brown.edu/__;!!CzAuKJ42GuquVTTmVmPViYEvSg!KsbhvmLlx8TkLK7y2NKz59hK4-4H7KXVV7dEyUG4vcQzi4Mh7nO-9HupA7_ep2V2p9KkD_i00tcg6nDqczDwOWf8YDMv$>

>>>>>

>>>>>

>>>>> <https://urldefense.com/v3/__https://sph.brown.edu/events/10-year-anniversary__;!!CzAuKJ42GuquVTTmVmPViYEvSg!KsbhvmLlx8TkLK7y2NKz59hK4-4H7KXVV7dEyUG4vcQzi4Mh7nO-9HupA7_ep2V2p9KkD_i00tcg6nDqczDwOcsy9Aer$>

>>>>>

>>>>>

>>>>>

>>>>> Aditya S. Khanna, Ph.D.

>>>>>

>>>>> Assistant Professor

>>>>>

>>>>> Department of Behavioral and Social Sciences

>>>>>

>>>>> Center for Alcohol and Addiction Studies

>>>>>

>>>>> Brown University School of Public Health

>>>>>

>>>>> Pronouns: he/him/his

>>>>>

>>>>>

>>>>> 401-863-6616

>>>>>

>>>>> sph.brown.edu

>>>>> <https://urldefense.com/v3/__https://sph.brown.edu/__;!!CzAuKJ42GuquVTTmVmPViYEvSg!KsbhvmLlx8TkLK7y2NKz59hK4-4H7KXVV7dEyUG4vcQzi4Mh7nO-9HupA7_ep2V2p9KkD_i00tcg6nDqczDwOWf8YDMv$>

>>>>>

>>>>> https://vivo.brown.edu/display/akhann16

>>>>> <https://urldefense.com/v3/__https://vivo.brown.edu/display/akhann16__;!!CzAuKJ42GuquVTTmVmPViYEvSg!KsbhvmLlx8TkLK7y2NKz59hK4-4H7KXVV7dEyUG4vcQzi4Mh7nO-9HupA7_ep2V2p9KkD_i00tcg6nDqczDwOWy55iTf$>

>>>>>

>>>>>

>>>>> _______________________________________________

>>>>> statnet_help mailing list

>>>>> statnet_help at u.washington.edu

>>>>> https://urldefense.com/v3/__http://mailman13.u.washington.edu/mailman/listinfo/statnet_help__;!!CzAuKJ42GuquVTTmVmPViYEvSg!KsbhvmLlx8TkLK7y2NKz59hK4-4H7KXVV7dEyUG4vcQzi4Mh7nO-9HupA7_ep2V2p9KkD_i00tcg6nDqczDwObRNh35k$

>>>> _______________________________________________

>>>> statnet_help mailing list

>>>> statnet_help at u.washington.edu

>>>> http://mailman13.u.washington.edu/mailman/listinfo/statnet_help

>>>> <https://urldefense.com/v3/__http://mailman13.u.washington.edu/mailman/listinfo/statnet_help__;!!CzAuKJ42GuquVTTmVmPViYEvSg!K2TPlppMmLqkp0AMD_Fj8vqeLS9FI7fJTx0UxQMy3qt_I1hNTinFaFOclmyytc3bbAXyEkzYTJhdrHccGTc77FsB$>

>>>>

>>> _______________________________________________

>>> statnet_help mailing list

>>> statnet_help at u.washington.edu

>>> http://mailman13.u.washington.edu/mailman/listinfo/statnet_help

>>> <https://urldefense.com/v3/__http://mailman13.u.washington.edu/mailman/listinfo/statnet_help__;!!CzAuKJ42GuquVTTmVmPViYEvSg!LKGctG3KpUbsSQwk8O2E14uWvIjFBhbEa_mHZ5i_MJ9vtXw_UfiYpxvp4p1HYdnN3aFwRdIYqyjPTzdZdoKq5yPB$>

>>>

>>

>> _______________________________________________

>> statnet_help mailing list

>> statnet_help at u.washington.edu

>> http://mailman13.u.washington.edu/mailman/listinfo/statnet_help

> --

> ***********************************************************************************************************************

> Steven M. Goodreau / Professor / Dept. of Anthropology / Adjunct Prof. / Dept. of Epidemiology

> (STEE-vun GOOD-roe) / he-him /https://faculty.washington.edu/goodreau / dzidzəlalič, x̌ʷəlč

> Physical address: Denny Hall M236; Mailing address: Campus Box 353100 / 4216 Memorial Way NE

> Univ. of Washington / Seattle WA 98195 / 1-206-685-3870 (phone) / 1-206-543-3285 (fax)

> "Fight for the things that you care about, but do it in a way that will lead others to join you" - Justice RB Ginsburg

> ***********************************************************************************************************************


--
***********************************************************************************************************************
Steven M. Goodreau / Professor / Dept. of Anthropology / Adjunct Prof. / Dept. of Epidemiology
(STEE-vun GOOD-roe) / he-him /https://faculty.washington.edu/goodreau / dzidzəlalič, x̌ʷəlč
Physical address: Denny Hall M236; Mailing address: Campus Box 353100 / 4216 Memorial Way NE
Univ. of Washington / Seattle WA 98195 / 1-206-685-3870 (phone) / 1-206-543-3285 (fax)
"Fight for the things that you care about, but do it in a way that will lead others to join you" - Justice RB Ginsburg
***********************************************************************************************************************
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://mailman13.u.washington.edu/pipermail/statnet_help/attachments/20241117/89f28f04/attachment-0001.html>


More information about the statnet_help mailing list