What is joint estimation?Variance-gamma distribution: parameter estimationSimilarities and differences...

How to execute a project with two resources where you need three resources?

Closest thing to Infinity Gauntlet in DnD5e

Little Endian Number to String Conversion

Chain with double bond or triple bond

Can I perform Umrah while on a Saudi Arabian visit e-visa

Always show full URL in Safari address bar

What is the best DIY approach to keeping brake dust off your rims?

How long should a test wait to assume that the result remains fixed

Can a successful economy exist without renting stuff?

5v home network

Should I withdraw my paper because the editor is delaying the report?

What is this plane with its thick cockpit?

Why is it so hard to land on The Moon?

What does IKEA-like mean?

SSH From a shared workplace computer

Did Feynman cite a fallacy about only circles having the same width in all directions as a reason for the Challenger disaster?

What is joint estimation?

Adding elements to some sublists of unequal length

How do you translate "Don't Fear the Reaper" into Latin?

D&D Monsters and Copyright

What kind of tools would be used to carve bone?

Does journal access significantly influence choice in which journal to publish in?

"a sign that one should be computing K-theory"

A Society Built Around Theft?



What is joint estimation?


Variance-gamma distribution: parameter estimationSimilarities and differences between regression and estimationCan someone explain the concept of nontrivial joint sufficient statistic for a pair of parameters?What location parameter is modelled by robust regression?What does it mean for a sample size to be “too low”, in terms of estimation vs inferrence?Logistic regression diagnostic plots in R






.everyoneloves__top-leaderboard:empty,.everyoneloves__mid-leaderboard:empty,.everyoneloves__bot-mid-leaderboard:empty{
margin-bottom:0;
}








6














$begingroup$


My question is simple as that: what is joint estimation? And what does it mean in the context of regression analysis? How is it done? I wandered in the mighty Internet for quite some time but did not find answers to these questions.










share|cite|improve this question









$endgroup$























    6














    $begingroup$


    My question is simple as that: what is joint estimation? And what does it mean in the context of regression analysis? How is it done? I wandered in the mighty Internet for quite some time but did not find answers to these questions.










    share|cite|improve this question









    $endgroup$



















      6












      6








      6


      1



      $begingroup$


      My question is simple as that: what is joint estimation? And what does it mean in the context of regression analysis? How is it done? I wandered in the mighty Internet for quite some time but did not find answers to these questions.










      share|cite|improve this question









      $endgroup$




      My question is simple as that: what is joint estimation? And what does it mean in the context of regression analysis? How is it done? I wandered in the mighty Internet for quite some time but did not find answers to these questions.







      regression estimation






      share|cite|improve this question













      share|cite|improve this question











      share|cite|improve this question




      share|cite|improve this question



      share|cite|improve this question










      asked 8 hours ago









      Lost in regressionLost in regression

      312 bronze badges




      312 bronze badges

























          3 Answers
          3






          active

          oldest

          votes


















          6
















          $begingroup$

          Joint estimation is, simply, jointly estimating two (or more) things at the same time. It can be as simple as estimating the mean and standard deviation from a sample.



          In a lot of the literature, the term is invoked because a special estimating procedure has to be used. This is usually the case when one quantity depends on the other and vice versa so that an analytic solution to the problem is intractable. How exactly joint estimation is done depends entirely on the problem.



          One method that pops up often for "joint modeling" or joint estimation is the EM-algorithm. EM stands for expectation - maximization. By alternating these steps, the E-step fills in the missing data that otherwise depend on component A, and the M-step finds optimal estimates for component B. By iterating the E and M steps, you can find a maximum likelihood estimate of A and B, thus jointly estimate these things.






          share|cite|improve this answer












          $endgroup$























            1
















            $begingroup$

            In a statistical context, the term "joint estimation" could conceivably mean one of two things:





              1. The simultaneous estimation of two scalar parameters (or equivalently, the estimation of a vector parameter with at least two elements); or



              1. The estimation of a single parameter pertaining to a joint (e.g., in the study of carpentry, plumbing systems, or marijuana).



            Of those two options, the second one is a joke, so almost certainly, joint estimation refers to simultaneously estimating two scalar parameters at once.






            share|cite|improve this answer










            $endgroup$























              0
















              $begingroup$

              Joint estimation is using data to estimate two or more parameters at the same time. Separate estimation evaluates each parameter one at a time.



              Estimation is the result of some form of optimization process. Because of this, there do not exist unique estimation solutions in statistics. If you change your goal, then you change what is optimal. When you first learn things such as regression, no one tells you why you are doing what you are doing. The goal of the instructor is to give you a degree of basic functionality using methods that work in a wide range of circumstances. At the beginning, you are not learning about regression. Instead, you are learning one or two regression methods that are widely applicable in a wide range of circumstances.



              The fact you are looking for solutions that solve a hidden goal makes it a bit difficult to understand.



              In the context of regression, imagine the following algebraic expression is true $$z=beta_xx+beta_yy+alpha$$. A truism in statistics is the more information that you have, the better off you are. Let us assume that you need to determine what values for $z$ will happen when you see $(x,y)$. The problem is that you do not know the true values for ${beta_x,beta_y,alpha}$. You have a large, complete data set of ${x,y,z}$.



              In separate estimation, you would estimate one parameter at a time. In joint estimation, you would estimate all of them at once.



              As a rule of thumb, joint estimation is more accurate than a separate estimate with a large complete data set. There is one general exception to that. Imagine you have a large set of $x$ and $z$ but a small set of $y$. Imagine most of your $y$ values are missing.



              In many estimation routines, you would delete the missing $x$s and $z$s and reduce down the set you are working from until all sets are complete. If you have deleted enough data, it can be more accurate to use the large number of $x$s and $z$s separately to estimate $z=beta_xx+alpha$ and $z=beta_yy+alpha$ than together.



              Now as to how it is done. All estimation, excluding a few exceptional cases, uses calculus to find an estimator that minimizes some form of loss or some type of risk. The concern is that you will be unlucky in choosing your sample. Unfortunately, there is an infinite number of loss functions. There is also an infinite number of risk functions.



              I found several videos for you because it is a giant topic so that you can look at it in a more general form. They are from Mathematical Monk.



              https://www.youtube.com/watch?v=6GhSiM0frIk



              https://www.youtube.com/watch?v=5SPm4TmYTX0



              https://www.youtube.com/watch?v=b1GxZdFN6cY



              and



              https://www.youtube.com/watch?v=WdnP1gmb8Hw.






              share|cite|improve this answer










              $endgroup$

















                Your Answer








                StackExchange.ready(function() {
                var channelOptions = {
                tags: "".split(" "),
                id: "65"
                };
                initTagRenderer("".split(" "), "".split(" "), channelOptions);

                StackExchange.using("externalEditor", function() {
                // Have to fire editor after snippets, if snippets enabled
                if (StackExchange.settings.snippets.snippetsEnabled) {
                StackExchange.using("snippets", function() {
                createEditor();
                });
                }
                else {
                createEditor();
                }
                });

                function createEditor() {
                StackExchange.prepareEditor({
                heartbeatType: 'answer',
                autoActivateHeartbeat: false,
                convertImagesToLinks: false,
                noModals: true,
                showLowRepImageUploadWarning: true,
                reputationToPostImages: null,
                bindNavPrevention: true,
                postfix: "",
                imageUploader: {
                brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
                contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/4.0/"u003ecc by-sa 4.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
                allowUrls: true
                },
                onDemand: true,
                discardSelector: ".discard-answer"
                ,immediatelyShowMarkdownHelp:true
                });


                }
                });















                draft saved

                draft discarded
















                StackExchange.ready(
                function () {
                StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f429712%2fwhat-is-joint-estimation%23new-answer', 'question_page');
                }
                );

                Post as a guest















                Required, but never shown

























                3 Answers
                3






                active

                oldest

                votes








                3 Answers
                3






                active

                oldest

                votes









                active

                oldest

                votes






                active

                oldest

                votes









                6
















                $begingroup$

                Joint estimation is, simply, jointly estimating two (or more) things at the same time. It can be as simple as estimating the mean and standard deviation from a sample.



                In a lot of the literature, the term is invoked because a special estimating procedure has to be used. This is usually the case when one quantity depends on the other and vice versa so that an analytic solution to the problem is intractable. How exactly joint estimation is done depends entirely on the problem.



                One method that pops up often for "joint modeling" or joint estimation is the EM-algorithm. EM stands for expectation - maximization. By alternating these steps, the E-step fills in the missing data that otherwise depend on component A, and the M-step finds optimal estimates for component B. By iterating the E and M steps, you can find a maximum likelihood estimate of A and B, thus jointly estimate these things.






                share|cite|improve this answer












                $endgroup$




















                  6
















                  $begingroup$

                  Joint estimation is, simply, jointly estimating two (or more) things at the same time. It can be as simple as estimating the mean and standard deviation from a sample.



                  In a lot of the literature, the term is invoked because a special estimating procedure has to be used. This is usually the case when one quantity depends on the other and vice versa so that an analytic solution to the problem is intractable. How exactly joint estimation is done depends entirely on the problem.



                  One method that pops up often for "joint modeling" or joint estimation is the EM-algorithm. EM stands for expectation - maximization. By alternating these steps, the E-step fills in the missing data that otherwise depend on component A, and the M-step finds optimal estimates for component B. By iterating the E and M steps, you can find a maximum likelihood estimate of A and B, thus jointly estimate these things.






                  share|cite|improve this answer












                  $endgroup$


















                    6














                    6










                    6







                    $begingroup$

                    Joint estimation is, simply, jointly estimating two (or more) things at the same time. It can be as simple as estimating the mean and standard deviation from a sample.



                    In a lot of the literature, the term is invoked because a special estimating procedure has to be used. This is usually the case when one quantity depends on the other and vice versa so that an analytic solution to the problem is intractable. How exactly joint estimation is done depends entirely on the problem.



                    One method that pops up often for "joint modeling" or joint estimation is the EM-algorithm. EM stands for expectation - maximization. By alternating these steps, the E-step fills in the missing data that otherwise depend on component A, and the M-step finds optimal estimates for component B. By iterating the E and M steps, you can find a maximum likelihood estimate of A and B, thus jointly estimate these things.






                    share|cite|improve this answer












                    $endgroup$



                    Joint estimation is, simply, jointly estimating two (or more) things at the same time. It can be as simple as estimating the mean and standard deviation from a sample.



                    In a lot of the literature, the term is invoked because a special estimating procedure has to be used. This is usually the case when one quantity depends on the other and vice versa so that an analytic solution to the problem is intractable. How exactly joint estimation is done depends entirely on the problem.



                    One method that pops up often for "joint modeling" or joint estimation is the EM-algorithm. EM stands for expectation - maximization. By alternating these steps, the E-step fills in the missing data that otherwise depend on component A, and the M-step finds optimal estimates for component B. By iterating the E and M steps, you can find a maximum likelihood estimate of A and B, thus jointly estimate these things.







                    share|cite|improve this answer















                    share|cite|improve this answer




                    share|cite|improve this answer



                    share|cite|improve this answer








                    edited 6 hours ago

























                    answered 8 hours ago









                    AdamOAdamO

                    38.8k2 gold badges71 silver badges156 bronze badges




                    38.8k2 gold badges71 silver badges156 bronze badges




























                        1
















                        $begingroup$

                        In a statistical context, the term "joint estimation" could conceivably mean one of two things:





                          1. The simultaneous estimation of two scalar parameters (or equivalently, the estimation of a vector parameter with at least two elements); or



                          1. The estimation of a single parameter pertaining to a joint (e.g., in the study of carpentry, plumbing systems, or marijuana).



                        Of those two options, the second one is a joke, so almost certainly, joint estimation refers to simultaneously estimating two scalar parameters at once.






                        share|cite|improve this answer










                        $endgroup$




















                          1
















                          $begingroup$

                          In a statistical context, the term "joint estimation" could conceivably mean one of two things:





                            1. The simultaneous estimation of two scalar parameters (or equivalently, the estimation of a vector parameter with at least two elements); or



                            1. The estimation of a single parameter pertaining to a joint (e.g., in the study of carpentry, plumbing systems, or marijuana).



                          Of those two options, the second one is a joke, so almost certainly, joint estimation refers to simultaneously estimating two scalar parameters at once.






                          share|cite|improve this answer










                          $endgroup$


















                            1














                            1










                            1







                            $begingroup$

                            In a statistical context, the term "joint estimation" could conceivably mean one of two things:





                              1. The simultaneous estimation of two scalar parameters (or equivalently, the estimation of a vector parameter with at least two elements); or



                              1. The estimation of a single parameter pertaining to a joint (e.g., in the study of carpentry, plumbing systems, or marijuana).



                            Of those two options, the second one is a joke, so almost certainly, joint estimation refers to simultaneously estimating two scalar parameters at once.






                            share|cite|improve this answer










                            $endgroup$



                            In a statistical context, the term "joint estimation" could conceivably mean one of two things:





                              1. The simultaneous estimation of two scalar parameters (or equivalently, the estimation of a vector parameter with at least two elements); or



                              1. The estimation of a single parameter pertaining to a joint (e.g., in the study of carpentry, plumbing systems, or marijuana).



                            Of those two options, the second one is a joke, so almost certainly, joint estimation refers to simultaneously estimating two scalar parameters at once.







                            share|cite|improve this answer













                            share|cite|improve this answer




                            share|cite|improve this answer



                            share|cite|improve this answer










                            answered 6 hours ago









                            BenBen

                            38.2k2 gold badges50 silver badges167 bronze badges




                            38.2k2 gold badges50 silver badges167 bronze badges


























                                0
















                                $begingroup$

                                Joint estimation is using data to estimate two or more parameters at the same time. Separate estimation evaluates each parameter one at a time.



                                Estimation is the result of some form of optimization process. Because of this, there do not exist unique estimation solutions in statistics. If you change your goal, then you change what is optimal. When you first learn things such as regression, no one tells you why you are doing what you are doing. The goal of the instructor is to give you a degree of basic functionality using methods that work in a wide range of circumstances. At the beginning, you are not learning about regression. Instead, you are learning one or two regression methods that are widely applicable in a wide range of circumstances.



                                The fact you are looking for solutions that solve a hidden goal makes it a bit difficult to understand.



                                In the context of regression, imagine the following algebraic expression is true $$z=beta_xx+beta_yy+alpha$$. A truism in statistics is the more information that you have, the better off you are. Let us assume that you need to determine what values for $z$ will happen when you see $(x,y)$. The problem is that you do not know the true values for ${beta_x,beta_y,alpha}$. You have a large, complete data set of ${x,y,z}$.



                                In separate estimation, you would estimate one parameter at a time. In joint estimation, you would estimate all of them at once.



                                As a rule of thumb, joint estimation is more accurate than a separate estimate with a large complete data set. There is one general exception to that. Imagine you have a large set of $x$ and $z$ but a small set of $y$. Imagine most of your $y$ values are missing.



                                In many estimation routines, you would delete the missing $x$s and $z$s and reduce down the set you are working from until all sets are complete. If you have deleted enough data, it can be more accurate to use the large number of $x$s and $z$s separately to estimate $z=beta_xx+alpha$ and $z=beta_yy+alpha$ than together.



                                Now as to how it is done. All estimation, excluding a few exceptional cases, uses calculus to find an estimator that minimizes some form of loss or some type of risk. The concern is that you will be unlucky in choosing your sample. Unfortunately, there is an infinite number of loss functions. There is also an infinite number of risk functions.



                                I found several videos for you because it is a giant topic so that you can look at it in a more general form. They are from Mathematical Monk.



                                https://www.youtube.com/watch?v=6GhSiM0frIk



                                https://www.youtube.com/watch?v=5SPm4TmYTX0



                                https://www.youtube.com/watch?v=b1GxZdFN6cY



                                and



                                https://www.youtube.com/watch?v=WdnP1gmb8Hw.






                                share|cite|improve this answer










                                $endgroup$




















                                  0
















                                  $begingroup$

                                  Joint estimation is using data to estimate two or more parameters at the same time. Separate estimation evaluates each parameter one at a time.



                                  Estimation is the result of some form of optimization process. Because of this, there do not exist unique estimation solutions in statistics. If you change your goal, then you change what is optimal. When you first learn things such as regression, no one tells you why you are doing what you are doing. The goal of the instructor is to give you a degree of basic functionality using methods that work in a wide range of circumstances. At the beginning, you are not learning about regression. Instead, you are learning one or two regression methods that are widely applicable in a wide range of circumstances.



                                  The fact you are looking for solutions that solve a hidden goal makes it a bit difficult to understand.



                                  In the context of regression, imagine the following algebraic expression is true $$z=beta_xx+beta_yy+alpha$$. A truism in statistics is the more information that you have, the better off you are. Let us assume that you need to determine what values for $z$ will happen when you see $(x,y)$. The problem is that you do not know the true values for ${beta_x,beta_y,alpha}$. You have a large, complete data set of ${x,y,z}$.



                                  In separate estimation, you would estimate one parameter at a time. In joint estimation, you would estimate all of them at once.



                                  As a rule of thumb, joint estimation is more accurate than a separate estimate with a large complete data set. There is one general exception to that. Imagine you have a large set of $x$ and $z$ but a small set of $y$. Imagine most of your $y$ values are missing.



                                  In many estimation routines, you would delete the missing $x$s and $z$s and reduce down the set you are working from until all sets are complete. If you have deleted enough data, it can be more accurate to use the large number of $x$s and $z$s separately to estimate $z=beta_xx+alpha$ and $z=beta_yy+alpha$ than together.



                                  Now as to how it is done. All estimation, excluding a few exceptional cases, uses calculus to find an estimator that minimizes some form of loss or some type of risk. The concern is that you will be unlucky in choosing your sample. Unfortunately, there is an infinite number of loss functions. There is also an infinite number of risk functions.



                                  I found several videos for you because it is a giant topic so that you can look at it in a more general form. They are from Mathematical Monk.



                                  https://www.youtube.com/watch?v=6GhSiM0frIk



                                  https://www.youtube.com/watch?v=5SPm4TmYTX0



                                  https://www.youtube.com/watch?v=b1GxZdFN6cY



                                  and



                                  https://www.youtube.com/watch?v=WdnP1gmb8Hw.






                                  share|cite|improve this answer










                                  $endgroup$


















                                    0














                                    0










                                    0







                                    $begingroup$

                                    Joint estimation is using data to estimate two or more parameters at the same time. Separate estimation evaluates each parameter one at a time.



                                    Estimation is the result of some form of optimization process. Because of this, there do not exist unique estimation solutions in statistics. If you change your goal, then you change what is optimal. When you first learn things such as regression, no one tells you why you are doing what you are doing. The goal of the instructor is to give you a degree of basic functionality using methods that work in a wide range of circumstances. At the beginning, you are not learning about regression. Instead, you are learning one or two regression methods that are widely applicable in a wide range of circumstances.



                                    The fact you are looking for solutions that solve a hidden goal makes it a bit difficult to understand.



                                    In the context of regression, imagine the following algebraic expression is true $$z=beta_xx+beta_yy+alpha$$. A truism in statistics is the more information that you have, the better off you are. Let us assume that you need to determine what values for $z$ will happen when you see $(x,y)$. The problem is that you do not know the true values for ${beta_x,beta_y,alpha}$. You have a large, complete data set of ${x,y,z}$.



                                    In separate estimation, you would estimate one parameter at a time. In joint estimation, you would estimate all of them at once.



                                    As a rule of thumb, joint estimation is more accurate than a separate estimate with a large complete data set. There is one general exception to that. Imagine you have a large set of $x$ and $z$ but a small set of $y$. Imagine most of your $y$ values are missing.



                                    In many estimation routines, you would delete the missing $x$s and $z$s and reduce down the set you are working from until all sets are complete. If you have deleted enough data, it can be more accurate to use the large number of $x$s and $z$s separately to estimate $z=beta_xx+alpha$ and $z=beta_yy+alpha$ than together.



                                    Now as to how it is done. All estimation, excluding a few exceptional cases, uses calculus to find an estimator that minimizes some form of loss or some type of risk. The concern is that you will be unlucky in choosing your sample. Unfortunately, there is an infinite number of loss functions. There is also an infinite number of risk functions.



                                    I found several videos for you because it is a giant topic so that you can look at it in a more general form. They are from Mathematical Monk.



                                    https://www.youtube.com/watch?v=6GhSiM0frIk



                                    https://www.youtube.com/watch?v=5SPm4TmYTX0



                                    https://www.youtube.com/watch?v=b1GxZdFN6cY



                                    and



                                    https://www.youtube.com/watch?v=WdnP1gmb8Hw.






                                    share|cite|improve this answer










                                    $endgroup$



                                    Joint estimation is using data to estimate two or more parameters at the same time. Separate estimation evaluates each parameter one at a time.



                                    Estimation is the result of some form of optimization process. Because of this, there do not exist unique estimation solutions in statistics. If you change your goal, then you change what is optimal. When you first learn things such as regression, no one tells you why you are doing what you are doing. The goal of the instructor is to give you a degree of basic functionality using methods that work in a wide range of circumstances. At the beginning, you are not learning about regression. Instead, you are learning one or two regression methods that are widely applicable in a wide range of circumstances.



                                    The fact you are looking for solutions that solve a hidden goal makes it a bit difficult to understand.



                                    In the context of regression, imagine the following algebraic expression is true $$z=beta_xx+beta_yy+alpha$$. A truism in statistics is the more information that you have, the better off you are. Let us assume that you need to determine what values for $z$ will happen when you see $(x,y)$. The problem is that you do not know the true values for ${beta_x,beta_y,alpha}$. You have a large, complete data set of ${x,y,z}$.



                                    In separate estimation, you would estimate one parameter at a time. In joint estimation, you would estimate all of them at once.



                                    As a rule of thumb, joint estimation is more accurate than a separate estimate with a large complete data set. There is one general exception to that. Imagine you have a large set of $x$ and $z$ but a small set of $y$. Imagine most of your $y$ values are missing.



                                    In many estimation routines, you would delete the missing $x$s and $z$s and reduce down the set you are working from until all sets are complete. If you have deleted enough data, it can be more accurate to use the large number of $x$s and $z$s separately to estimate $z=beta_xx+alpha$ and $z=beta_yy+alpha$ than together.



                                    Now as to how it is done. All estimation, excluding a few exceptional cases, uses calculus to find an estimator that minimizes some form of loss or some type of risk. The concern is that you will be unlucky in choosing your sample. Unfortunately, there is an infinite number of loss functions. There is also an infinite number of risk functions.



                                    I found several videos for you because it is a giant topic so that you can look at it in a more general form. They are from Mathematical Monk.



                                    https://www.youtube.com/watch?v=6GhSiM0frIk



                                    https://www.youtube.com/watch?v=5SPm4TmYTX0



                                    https://www.youtube.com/watch?v=b1GxZdFN6cY



                                    and



                                    https://www.youtube.com/watch?v=WdnP1gmb8Hw.







                                    share|cite|improve this answer













                                    share|cite|improve this answer




                                    share|cite|improve this answer



                                    share|cite|improve this answer










                                    answered 7 hours ago









                                    Dave HarrisDave Harris

                                    4,3905 silver badges17 bronze badges




                                    4,3905 silver badges17 bronze badges


































                                        draft saved

                                        draft discarded



















































                                        Thanks for contributing an answer to Cross Validated!


                                        • Please be sure to answer the question. Provide details and share your research!

                                        But avoid



                                        • Asking for help, clarification, or responding to other answers.

                                        • Making statements based on opinion; back them up with references or personal experience.


                                        Use MathJax to format equations. MathJax reference.


                                        To learn more, see our tips on writing great answers.




                                        draft saved


                                        draft discarded














                                        StackExchange.ready(
                                        function () {
                                        StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f429712%2fwhat-is-joint-estimation%23new-answer', 'question_page');
                                        }
                                        );

                                        Post as a guest















                                        Required, but never shown





















































                                        Required, but never shown














                                        Required, but never shown












                                        Required, but never shown







                                        Required, but never shown

































                                        Required, but never shown














                                        Required, but never shown












                                        Required, but never shown







                                        Required, but never shown







                                        Popular posts from this blog

                                        Taj Mahal Inhaltsverzeichnis Aufbau | Geschichte | 350-Jahr-Feier | Heutige Bedeutung | Siehe auch |...

                                        Baia Sprie Cuprins Etimologie | Istorie | Demografie | Politică și administrație | Arii naturale...

                                        Nicolae Petrescu-Găină Cuprins Biografie | Opera | In memoriam | Varia | Controverse, incertitudini...